Unnamed: 0
int64 0
7.24k
| id
int64 1
7.28k
| raw_text
stringlengths 9
124k
| vw_text
stringlengths 12
15k
|
---|---|---|---|
7,200 | 961 | Transformation Invariant Autoassociation
with Application to
Handwritten Character Recognition
Holger Schwenk
Maurice Milgram
PARC
Universite Pierre et Marie Curie
tour 66-56, boite 164
4, place Jussieu, 75252 Paris cedex 05, France.
e-mail: schwenk@robo.jussieu.fr
Abstract
When training neural networks by the classical backpropagation algorithm the whole problem to learn must be expressed by a set of inputs and
desired outputs. However, we often have high-level knowledge about
the learning problem. In optical character recognition (OCR), for instance, we know that the classification should be invariant under a set of
transformations like rotation or translation. We propose a new modular
classification system based on several autoassociative multilayer perceptrons which allows the efficient incorporation of such knowledge. Results
are reported on the NIST database of upper case handwritten letters and
compared to other approaches to the invariance problem.
1 INCORPORATION OF EXPLICIT KNOWLEDGE
The aim of supervised learning is to learn a mapping between the input and the output
space from a set of example pairs (input, desired output). The classical implementation
in the domain of neural networks is the backpropagation algorithm. If this learning set is
sufficiently representative of the underlying data distributions, one hopes that after learning,
the system is able to generalize correctly to other inputs of the same distribution.
992
Holger Schwenk, Maurice Milgram
It would be better to have more powerful techniques to incorporate knowledge into the
learning process than the choice of a set of examples. The use of additional knowledge
is often limited to the feature extraction module. Besides simple operations like (size)
normalization, we can find more sophisticated approaches like zernike moments in the
domain of optical character recognition (OCR). In this paper we will not investigate this
possibility, all discussed classifiers work directly on almost non preprocessed data (pixels).
In the context of OCR interest focuses on invariance of the classifier under a number of
given transformations (translation, rotation, ... ) of the data to classify. In general a neural
network could extract those properties of a large enough learning set, but it is very hard to
learn and will probably take a lot of time. In the last years two main approaches for this
invariance problem have been proposed: tangent-prop and tangent-distance. An indirect
incorporation can be achieved by boosting (Drucker, Schapire and Simard, 1993).
In this paper we briefly discuss these approaches and will present a new classification system
which allows the efficient incorporation of transformation invariances.
1.1 TANGENT PROPAGATION
The principle of tangent-prop is to specify besides desired outputs also desired changes jJJ.
of the output vector when transforming the net input x by the transformations tJJ. (Simard,
Victorri, LeCun and Denker, 1992).
For this, let us define a transformation of pattern p as t(p, a) : P --t P where P is the space
of all patterns and a a parameter. Such transformations are in general highly nonlinear
operations in the pixel space P and their analytical expressions are seldom known. It is
therefore favorable to use a first order approximation:
t(p, a) :::::: p
+ atp
with
- at(p, a)
tp _
aa
I
(1)
a=O
tp is called the tangent vector. This definition can be generalized to c transformations:
t(p, Q) :::::: p + a] tp] + ... + a c tpc = P + TpQ
where Tp is a n x c matrix, each column corresponding to a tangent vector.
(2)
Let us define R (x) the function calculated by the network. The desired behavior of the net
outputs can be obtained by adding a regularization term Er to the objective function:
E,
=
~
k -aR(t;~.a)tJ '" ~ k -a~~x) t~ I '
(3)
ti
is the tangent vector for transformation tJJ. of the input vector x and a R (x) / ax is the
gradient of the network with respect to the inputs. Transformation invariance of the outputs
is obtained by setting jJJ. = 0, so we want that aR (x) / ax is orthogonal to ti .
Tangent-prop improved the learning time and the generalization on small databases, but its
applicability to highly constraint networks (many shared weights) trained on large databases
remains unknown.
1.2 TANGENT DISTANCE
Another class of classifiers are memory based learning methods which rely on distance
metrics. The incorporation of knowledge in such classifiers can be done by a distance
Transformation Invariant Autoassociation
993
measure which is (locally) invariant under a set of specified transformations.
(Simard, LeCun and Denker, 1993) define tangent distance as the minimal distance between
the two hyperplanes spanned up by the tangent vectors Tp in point P and Tq in point q:
Dpq(p, q)
= ~i!J (p + Tpii 01,/3
q - TqiJ) 2
= (p + Tpii*
_ q _ TqiJ*) 2
(4)
The optimality condition is that the partial derivatives 8Dpq /8ii* and 8D pq /8iJ* should
be zero. The values ii* and iJ* minimizing (4) can be computed by solving these two linear
systems numerically.
(Simard, LeCun and Denker, 1993) obtained very good results on handwritten digits and
letters using tangent distance with a I-nearest-neighborclassifier (I-nn) . A big problem of
every nn-classifier, however, is that it uses no compilation of the data and it needs therefore
numerous reference vectors resulting in long classification time and high memory usage.
Like reported in (Simard, 1994) and (Sperdutti and Stork, 1995) important improvements
are possible, but often a trade-off between speed and memory usage must be made.
2
ARCHITECTURE OF THE CLASSIFIER
The main idea of our approach is to use an autoassociative multilayer perceptron with a
low dimensional hidden layer for each class to recognize. These networks, called diabolo
network in the following, are trained only with examples of the corresponding class. This
can be seen as supervised learning without counter-examples. Each network learns a hidden
layer representation which preserves optimally the information of the examples of one class.
These learned networks can now be used like discriminant functions: the reconstruction
error is in general much lower for examples of the learned class than for the other ones.
In order to build a classifier we use a decision module which interprets the distances between
the reconstructed output vectors and the presented example. In our studies we have used
until now a simple minimum operator which associates the class of the net with the smallest
distance (Fig. 1).
The figure illustrates also typical classification behavior, here when presenting a "D" out of
the test set. One can see clearly that the distance of the network "D" is much lower than for
the two other ones. The character is therefore correctly classified. It is also interesting to
analyze the outputs of the two networks with the next nearest distances: the network "0"
tries to output a more round character and the network "B" wants to add a horizontal bar in
the middle.
The basic classification architecture can be adapted in two ways to the problem to be solved.
One on hand we can imagine different architectures for each diabolo network, e.g. several
encoding/decoding layers which allow nonlinear dimension reduction. It is even possible
to use shared weights realizing local feature detectors (see (Schwenk and Milgram, 1994)
for more details).
One the other hand we can change the underlying distance measure, as long as the derivatives
with respect to the weights can be calculated. This offers a powerful and efficient mechanism
to introduce explicit knowledge into the learning algorithm of a neural network. In the
discussed case, the recognition of characters represented as pixel images, we can use a
994
Holger Schwenk, Maurice Milgram
score B
8.07
I--t-~
A
U.
scoreD
--~
4.49
score 0
8.54
character
to classify
input
vector
diabolo
networks
output
vectors
distance
measures
decision
module
Figure 1: Basic Architecture of a Diabolo Classifier
transformation invariant distance measure between the net output 0 and the desired output
d (that is of course identical with the net input). The networks do not need to learn each
example separately any more, but they can use the set of specified transformations in order
to find a common non linear model of each class.
The advantage of this approach, besides a better expected generalization behavior of course,
is a very low additional complexity. In comparison to the original k-nn approach, and
supposedly any possible optimization, we need to calculate only one distance measure for
each class to recognize, regardless of the number of learning data.
We used two different versions of tangent distance with increasing complexity:
1. single sided tangent distance:
Dd(d,o)
=mjn~
(d+Td a - o
f =~ (
d+Td a *
-of
(5)
This is the minimal distance between the hyperplane spanned up by the tangent
vectors Td in input vector d and the untransformed output vector o.
2. double sided tangent distance:
Ddo(d,o) = mi!?! (d
Ci,/3 2
+ Tda -
0
*g -
ToiJ)
2
(6)
The convolution of the net output with a Gaussian 9 is necessary for the computation of the tangent vectors To (the net input d is convolved during preprocessing).
Figure 2 shows a graphical comparison of Euclidean distance with the two tangent distances.
Transformation Invariant Autoassociation
995
d : desired output
td : tangent vector in d
o : net output
to : tangent vector in 0
D : Euclidean distance
D d : single sided tangent distance
(only d is transformed)
Ddo : double sided tangent distance
(both points are transformed)
V' D d : gradient of D d
Figure 2: Comparison of Euclidean Distance with the Different Tangent Distances
The major advantage of the single sided version is that we can now calculate easily the
optimal multipliers &* and therefore the whole distance (the double sided version demands
expensive matrix multiplications and the numerical solution of a system of linear equations).
The optimality condition 8Dd(d, 0)/8&* ~ OT gives:
(7)
The tangent vectors Td and the matrix Ti} = (TITd)-1 can be precomputed and stored in
memory. Note that it is the same for all diabolo networks, regardless of their class.
2.1
LEARNING ALGORITHM
When using a tangent distance with an autoencoder we must calculate its derivatives with
respect to the weights, i.e. after application of the chain rule with respect to the output
vector o. In the case of the single sided tangent distance we get:
(8)
The resulting learning algorithm is therefore barely more complicated than with standard
Euclidean error. Furthermore it has a pleasant graphical interpretation: the net output
doesn't approach directly the desired output any more, but it takes the shortest way towards
the tangent hyperplane (see also fig. 2).
The derivation of the double sided tangent distance with respect to the net output is more
complicated. In particular we must derivate the convolution of the net output with a
Gaussian as well as the tangent vectors To. These equations will be published elsewhere.
Training of the whole system is stopped when the error on the cross validation set reaches a
minimum. Using stochastic gradient descent convergence is typically achieved after some
ten iterations.
996
3
Holger Schwenk, Maurice Milgram
APPLICATION TO CHARACTER RECOGNITION
In 1992 the National Institute of Standards and Technology provided a Database of handwritten digits and letters, known under the name NIST Special-Database 3. This database
contains about 45000 upper case segmented characters which we have divided into learning
and cross-validation set (60%) and test set (40%) respectively.
We only applied a very simple preprocessing: the binary characters were centered and sizenormalized (the aspect-ratio was kept). The net input is 16 x 16 pixels with real-values.
3.1 EXPERIMENTAL RESULTS
All the following results were obtained by fully connected diabolo networks with one low
dimensional hidden layer, and a set of eight transformations (x- and y-translation, rotation,
scaling, axis-deformation, diagonal-deformation, x- and y-thickness). Figure 3 illustrates
how the networks use the transformations.
Euclidean distance: 11 .1
,
,
input
output
Euclidean distance: 20.0
,
,
input
output
optimally ttansformed
input
output
optimally ttansformed
input
output
'"
tangent distance:
0.61
tangent distance: 0.94
.,
Figure 3: Reconstruction Examples (test set). The left side of each screen dump depicts
the input character and the right side the one reconstructed by the network. In the middle,
finally, one can see the optimally transformed patterns as calculated when evaluating the
double sided tangent distance, i.e. transformed by a* and jJ* respectively.
Although the "I.:' in the first example has an unusual short horizontal line, the network
reconstructs a normally sized character. It is clearly visible how the input transformation
lengthens and the output transformation shortens this line in order to get a small tangent
distance. The right side shows a very difficult classification problem: a heavily deformed
"T". Nevertheless we get a small tangent distance, so that the character is correctly classified.
In summary we note a big difference between the Euclidean and the tangent distances, this
is a good indicator that the networks really use the transformations.
The performances on the whole test set of about 18 000 characters are summarized in
figure 4. For comparison we give also the results of a one nearest neighbor classifier on
the same test set. The incorporation of knowledge improved in both cases dramatically
the performances. The diabolo classifier, for instance, achieves an error rate of 4.7 % with
simple Euclidean distance which goes down to 3.7 % with the single sided and to only
2.6 % with the double sided tangent distance. In order to get the same results with the
I-nn approach, the whole set of 27000 reference vectors had to be used. It's worth to note
the results with less references: when using only 18 000 reference vectors the error rates
increased to 3.7% for the single sided and to 2.8% for the double sided version respectively.
Transformation Invariant Autoassociation
%
997
5.3
4.7
r;;
5-
-
11
F
4.0
3.7
c
Euclidean
-
one sided
-
two sided
2.5
I-nn
Diabolo
LeNet
(27 000 refs)
Figure 4: Raw Error Rate with NIST Upper Case Letters (test set)
In practical applications we are not only interested in low error rates, but we need also low
computationally costs. An important factor is the recognition speed. The overall processing
time of a diabolo classifier using the full tangent distance corresponds to the calculation of
about 7 000 Euclidean and to less than 50 tangent distances. This should be less than for
any algorithm of the k-nn family. If we assume the precalculation of all the tangent vectors
and other expensive matrix multiplications, we could evaluate about 80 tangent distances,
but the price would be exploding memory requirements. A diabolo classifier, on the other
hand, needs only few memory: the storage of the weights corresponds to about 60 reference
vectors per class. On a HP 715/50 workstation we obtained a recognition rate of 7.5 chis
with the single sided and of more than 2.5 chis with the double sided tangent distance. We
have also a method to combine both by rejection, resulting in up to 4 chis at the same low
error rates (corresponds to the calculation of 32 double sided tangent distances).
The table contains also the results of a large multilayer perceptron with extensive use of
shared weights, known as LeNet. (Drucker, Schapire and Simard, 1993) give an error rate
of 4.0% when used alone and of 2.4% for an ensemble of three such networks trained
by boosting. The networks were trained on a basic set of 10 000 examples, the cross
validation and test set consisted of 2 000 and 3 000 examples respectively (Drucker,
personal communication). Due to the different number of examples, the results are perhaps
not exactly comparable, but we can deduce nevertheless that the state of the art on this
database seems to be around 2.5 %.
4 DISCUSSION
We have proposed a new classification architecture that allows the efficient incorporation of
knowledge into the learning algorithm. The system is easy to train and only one structural
parameter must be chosen by the supervisor: the size of the hidden layer. It achieved state
of the art recognition rates on the NIST database of handwritten upper case letters at a very
low computational complexity.
Furthermore we can say that a hardware implementation seems to be promising. Fully
connected networks with only two layers are easy to put into standardized hardware chips.
We could even propagate all diabolo networks in parallel. Speedups of several orders of
magnitude should therefore be possible.
998
Holger Schwenk, Maurice Milgram
On this year NIPS conference several authors presented related approaches. A comparable
classification architecture was proposed by (Hinton, Revow and Dayan, 1995). Instead of
one non-linear global model per class, several local linear models were used by performing
separately principal component analysis (PCA) on subsets of each class. Since diabolo
networks with one hidden layer and linear activation functions perform PCA, this architecture can be interpreted as an hierarchical diabolo classifier with linear nets and Euclidean
distance. Such an hierarchisation could also be done with our classifier, i.e. with tangent
distance and sigmoidal units, and might improve the results even further.
(Hastie, Simard and Sackinger, 1995) developed an iterative algorithm that learns optimal
reference vectors in the sense of tangent distance. An extension allows also to learn typical
invariant transformations, i.e. tangent vectors, of each class. These two algorithms allowed
to reduce drastically the number of reference vectors, but the results of the original approach
couldn't be achieved no longer.
Acknowledgements
The first author is supported by the German Academic Exchange Service under grant
HSP II 516.006.512.3. The simulations were performed with the Aspirin/ MIGRAINES
neural network simulator developed by the MITRE Corporation.
References
H. Drucker, R. Schapire, and P. Simard (1993), "Boosting performance in neural networks,"
Int. Journal of Pattern Recognition and Artificial Intelligence, vol. 7, no. 4, pp. 705-719.
T. Hastie, P. Simard and E. Sackinger (1995), "Learning prototype models for tangent
distance," in NIPS 7 (G. Tesauro, D. Touretzky, and T. Leen, eds.), Morgan Kaufmann.
G. Hinton, M. Revow, and P. Dayan (1995), "Recognizing handwritten digits using mixtures
of linear models," in NIPS 7 (G. Tesauro, D. Touretzky, and T. Leen, eds.), Morgan
Kaufmann.
H. Schwenk and M. Milgram (1994), "Structured diabolo-networks for handwritten character recognition," in International Conference on Artificial Neural Networks, pp. 985-988,
Springer-Verlag.
P. Simard, B. Victorri, Y. LeCun, and 1. Denker (1992), "Tangent prop - a formalism for
specifying selected invariances in an adaptive network," in NIPS 4 (1. Moody, S. Hanson,
and R. Lippmann, eds.), pp. 895-903, Morgan Kaufmann.
P. Simard, Y. LeCun, and J. Denker (1993), "Efficient pattern recognition using a new
transformation distance," in NIPS 5 (S. Hanson, J. Cowan, and C. Giles, eds.), pp. 50-58,
Morgan Kaufmann.
P. Simard (1994), "Efficient Computation of complex distance measures using hierarchical
filtering," in NIPS 6 (J.D. Cowan, G. Tesauro, and J. Alspector, eds.), pp. 50-58, Morgan
Kaufmann.
A. Sperdutti and D.G. Stork (1995), ''A rapid graph-based method for arbitrary transformation invariant pattern classification," in NIPS 7 (G. Tesauro, D. Touretzky, and T. Leen,
eds.), Morgan Kaufmann.
| 961 |@word deformed:1 version:4 briefly:1 middle:2 seems:2 simulation:1 propagate:1 reduction:1 moment:1 contains:2 score:2 activation:1 must:5 visible:1 numerical:1 alone:1 intelligence:1 selected:1 realizing:1 short:1 boosting:3 hyperplanes:1 sigmoidal:1 combine:1 introduce:1 expected:1 rapid:1 alspector:1 behavior:3 simulator:1 chi:3 td:5 increasing:1 provided:1 underlying:2 interpreted:1 developed:2 transformation:24 corporation:1 every:1 ti:3 exactly:1 classifier:14 normally:1 unit:1 grant:1 service:1 local:2 encoding:1 untransformed:1 might:1 specifying:1 autoassociation:4 limited:1 practical:1 lecun:5 backpropagation:2 digit:3 get:4 operator:1 storage:1 context:1 put:1 go:1 regardless:2 rule:1 spanned:2 imagine:1 heavily:1 us:1 associate:1 recognition:11 expensive:2 database:8 module:3 solved:1 calculate:3 connected:2 trade:1 counter:1 supposedly:1 transforming:1 complexity:3 personal:1 trained:4 solving:1 parc:1 easily:1 indirect:1 chip:1 schwenk:8 represented:1 derivation:1 train:1 artificial:2 couldn:1 modular:1 say:1 advantage:2 net:13 analytical:1 propose:1 reconstruction:2 fr:1 dpq:2 tpc:1 convergence:1 double:9 requirement:1 ij:2 nearest:3 lengthens:1 stochastic:1 centered:1 exchange:1 generalization:2 really:1 diabolo:14 extension:1 sufficiently:1 around:1 mapping:1 major:1 achieves:1 smallest:1 favorable:1 hope:1 clearly:2 gaussian:2 aim:1 zernike:1 ax:2 focus:1 improvement:1 sizenormalized:1 sense:1 dayan:2 nn:6 typically:1 hidden:5 transformed:4 france:1 interested:1 pixel:4 overall:1 classification:10 art:2 special:1 extraction:1 identical:1 holger:5 few:1 mjn:1 preserve:1 recognize:2 national:1 tq:1 interest:1 aspirin:1 investigate:1 possibility:1 highly:2 mixture:1 tj:1 compilation:1 chain:1 partial:1 necessary:1 orthogonal:1 euclidean:11 desired:8 deformation:2 minimal:2 stopped:1 instance:2 classify:2 column:1 increased:1 formalism:1 giles:1 ar:2 tp:5 applicability:1 cost:1 subset:1 tour:1 recognizing:1 supervisor:1 optimally:4 reported:2 stored:1 thickness:1 international:1 off:1 decoding:1 moody:1 reconstructs:1 maurice:5 simard:12 derivative:3 summarized:1 int:1 performed:1 try:1 lot:1 analyze:1 complicated:2 parallel:1 curie:1 kaufmann:6 ensemble:1 generalize:1 handwritten:7 raw:1 worth:1 published:1 classified:2 detector:1 reach:1 touretzky:3 milgram:7 tpq:1 definition:1 ed:6 pp:5 universite:1 mi:1 workstation:1 knowledge:9 sophisticated:1 supervised:2 specify:1 improved:2 leen:3 done:2 furthermore:2 until:1 hand:3 horizontal:2 sackinger:2 nonlinear:2 propagation:1 perhaps:1 name:1 usage:2 consisted:1 multiplier:1 regularization:1 lenet:2 round:1 during:1 generalized:1 presenting:1 image:1 common:1 rotation:3 stork:2 discussed:2 interpretation:1 numerically:1 tda:1 atp:1 seldom:1 hp:1 had:1 pq:1 longer:1 deduce:1 add:1 tesauro:4 verlag:1 binary:1 seen:1 minimum:2 additional:2 morgan:6 shortest:1 exploding:1 ii:3 full:1 segmented:1 academic:1 calculation:2 offer:1 long:2 cross:3 divided:1 basic:3 multilayer:3 metric:1 iteration:1 normalization:1 achieved:4 want:2 separately:2 derivate:1 victorri:2 ot:1 probably:1 cedex:1 cowan:2 structural:1 enough:1 easy:2 architecture:7 hastie:2 interprets:1 reduce:1 idea:1 prototype:1 ddo:2 drucker:4 expression:1 pca:2 jj:1 autoassociative:2 dramatically:1 pleasant:1 locally:1 ten:1 hardware:2 schapire:3 correctly:3 per:2 vol:1 nevertheless:2 preprocessed:1 marie:1 kept:1 graph:1 year:2 letter:5 powerful:2 place:1 almost:1 family:1 decision:2 scaling:1 comparable:2 layer:7 adapted:1 incorporation:7 constraint:1 aspect:1 speed:2 optimality:2 jjj:2 performing:1 optical:2 speedup:1 structured:1 character:15 invariant:9 sided:18 computationally:1 equation:2 remains:1 discus:1 precomputed:1 mechanism:1 german:1 know:1 unusual:1 operation:2 denker:5 eight:1 ocr:3 hierarchical:2 pierre:1 convolved:1 original:2 standardized:1 graphical:2 build:1 classical:2 objective:1 diagonal:1 gradient:3 distance:51 hsp:1 mail:1 discriminant:1 barely:1 besides:3 ratio:1 robo:1 minimizing:1 difficult:1 implementation:2 unknown:1 perform:1 upper:4 convolution:2 nist:4 descent:1 hinton:2 communication:1 arbitrary:1 pair:1 paris:1 specified:2 extensive:1 hanson:2 learned:2 nip:7 able:1 bar:1 pattern:6 memory:6 rely:1 indicator:1 improve:1 technology:1 numerous:1 axis:1 migraine:1 extract:1 autoencoder:1 acknowledgement:1 tangent:47 multiplication:2 fully:2 tjj:2 interesting:1 filtering:1 validation:3 principle:1 dd:2 translation:3 course:2 elsewhere:1 summary:1 supported:1 last:1 drastically:1 side:3 allow:1 perceptron:2 institute:1 neighbor:1 shortens:1 calculated:3 dimension:1 evaluating:1 doesn:1 author:2 made:1 adaptive:1 preprocessing:2 reconstructed:2 lippmann:1 global:1 iterative:1 table:1 promising:1 learn:5 complex:1 mitre:1 domain:2 main:2 whole:5 big:2 scored:1 allowed:1 ref:1 fig:2 representative:1 dump:1 screen:1 depicts:1 explicit:2 jussieu:2 learns:2 down:1 er:1 adding:1 ci:1 magnitude:1 illustrates:2 demand:1 rejection:1 expressed:1 springer:1 aa:1 corresponds:3 prop:4 sized:1 towards:1 shared:3 price:1 revow:2 hard:1 change:2 typical:2 hyperplane:2 principal:1 called:2 invariance:6 experimental:1 perceptrons:1 incorporate:1 evaluate:1 |
7,201 | 962 | Recognizing Handwritten Digits Using Mixtures of Linear Models
Geoffrey E Hinton
Michael Revow Peter Dayan
Deparbnent of Computer Science, University of Toronto
Toronto, Ontario, Canada M5S lA4
Abstract
We construct a mixture of locally linear generative models of a collection of pixel-based images of digits, and use them for recognition. Different models of a given digit are used to capture different
styles of writing, and new images are classified by evaluating their
log-likelihoods under each model. We use an EM-based algorithm
in which the M-step is computationally straightforward principal
components analysis (PCA). Incorporating tangent-plane information [12] about expected local deformations only requires adding
tangent vectors into the sample covariance matrices for the PCA,
and it demonstrably improves performance.
1 Introduction
The usual way of using a neural network for digit recognition is to train it to output
one of the ten classes. When the training data is limited to N examples equally
distributed among the classes, there are only N log2 10 bits of constraint in the class
labels so the number of free parameters that can be allowed in a discriminative
neural net model is severely limited. An alternative approach, motivated by density
estimation, is to train a separate autoencoder network on examples of each digit
class and to recognise digits by seeing which autoencoder network gives the best
reconstruction of the data. Because the output of each autoencoder has the same
dimensionality as the image, each training example provides many more bits of
constraint on the parameters. In the example we describe, 7000 training images are
sufficient to fit 384, 000 parameters and the training procedure is fast enough to do
the fitting overnight on an R4400-based machine.
Auto-encoders can be viewed in terms of minimum description length descriptions
of data in which the input-hidden weights produce a code for a particular case and
1016
Geoffrey E. Hinton, Michael Revow, Peter Dayan
the hidden-output weights embody a generative model which turns this code back
into a close approximation of the original example [14,7]. Code costs (under some
prior) and reconstruction error (squared error assuming an isotropic Gaussian
misfit model) sum to give the overall code length which can be viewed as a lower
bound on the log probability density that the autoencoder assigns to the image.
This properly places the emphasis on designing appropriate generative models
for the data which will generalise well by assigning high log probabilities to new
patterns from the same classes.
We apply this idea to recognising handwritten digits from grey-level pixel images using linear auto-encoders. Linear hidden units for autoencoders are barely
worse than non-linear ones when squared reconstruction error is used [I], but have
the great computational advantage during training that input-hidden and hiddenoutput weights can be derived from principal components analysis (PCA) of the
training data. In effect a PCA encoder approximates the entire N dimensional distribution of the data with a lower dimensional"Gaussian pancake" [13], choosing,
for optimal data compression, to retain just a few of the PCs. One could build a
single PCA model for each digit - however the many different styles of writing
suggest that more than one Gaussian pancake should be used, by dividing the
population of a single digit class into a number of sub-classes and approximating
each class by its own model. A similar idea for data compression was used by [9]
where vector quantization was used to define sub-classes and PCA was performed
within each sub-class (see also [2]). We used an iterative method based on the
Expectation Maximisation (EM) algorithm [4] to fit mixtures of linear models.
The reductio of the local linear approach would have just one training pattern in
each model. This approach would amount to a nearest neighbour method for
recognition using a Euclidean metric for error, a technique which is known to be
infelicitous. One reason for this poor performance is that characters do not only
differ in writing styles, but also can undergo simple transformations such as translations and rotations. These give rise to somewhat non-linear changes as measured
in pixel space. Nearest neighbour methods were dramatically improved methods
by defining a metric in which locally linearised versions of these transformations
cost nothing [12]. In their tangent distance method, each training or test point is
represented by a h-dimensionallinear subspace, where each of these dimensions
corresponds to a linear version of one of the transformations, and distances are
measured between subspaces rather than points. The local linear autoencoder
method can be seen just like this - variations along one of the h principal component directions are free, while variations along the remaining principal directions
cost. However, rather than storing and testing each training pattern, the local
models summarise the regularities over a number of patterns. This reduces storage and recognition time, and allows the directions of free variation to be averaged
over numbers of patterns and also to be determined by the data rather than being
pre-specified. A priori knowledge that particular transformations are important
can be incorporated using a version of the tangent-prop procedure [12], which is
equivalent in this case to adding in slightly transformed versions of the patterns.
Also reconstruction error could be assessed either at a test pattern, or, more like
tangent distance, between its transformation subspace and the models' principal
components subs paces.
Recognizing Handwritten Digits Using Mixtures of Linear Models
1017
,a
,,
,
Figure 1: Didactic example of tangent information and local linear models. See
text for details.
Figure 1 illustrates the idea. Imagine that the four points 1-4 portray in image
space different examples of the same digit, subject to some smooth transformation.
As in tangent distance, one could represent this curve using the points and their
local tangents (thick lines). However one might do better splitting it into three
local linear models rather than four - model 'a' Gust a line in this simple case)
averages the upper part of the curvp more effectively than the combination of the
two tangents at '1' and '2'. However, given just the points, one might c~nstruct
model 'b' for '3' and '4', which would be unfortunate. Incorporating information
about the tangents as well would encourage the separation of these segments. Care
should be taken in generalising this picture to high dimensional spaces.
The next section develops the theory behind variants of these systems (which is
very similar to that in [5, 10]), and section 3 discusses how they perform.
2 Theory
Linear auto-encoders embody a model in which variations from the mean of a
population along certain directions are cheaper than along others, as measured by
the log-unlikelihood of examples. Creating such a generative model is straightforward. Principal component analysis (PCA) is performed on the training data
and the leading h principal components are retained, defining an h-dimensional
subspace onto which the n-dimensional inputs are projected. We choose h using
cross-validation, although a minimum description length criterion could also be
used. We ignore the effect of different variances along the different principal components and use a model in which the code length for example i (the negative
log-likelihood) is proportional to the reconstruction error - the squared Euclidean
distance between the output of the autoencoder and the pattern itself.
Rather than having just one autoencoder for each digit, there is a whole collection,
and therefore we use EM to assign examples to n sub-classes, just as in clustering
using a mixture of Gaussians generative model. During the E-step, the responsibility for each pattern is assigned amongst the sub-classes, and in the M-step PCA
is performed, altering the parameters of a sub-class appropriately to minimise the
1018
Geoffrey E. Hinton, Michael Revow, Peter Dayan
reconstruction cost of the data for which it is responsible.
Formally, the algorithm is:
1. Choose initial autoencoder assignments for each example in the training
set (typically using a K-means clustering algorithm).
2. Perform PCA separately for each autoencoder;
3. Reassign patterns to the autoencoder that reconstructs them the best;
4. Stop if no patterns have changed sub-class, otherwise return to step 2.
There is a 'soft' version of the algorithm in which the responsibility of autoencoder
q for example i is calculated as ri.q = e-IIE,q 112 /Za 2 /(L.r e-IIE'TII2 /Z( 2 ) where Ei.r
is the reconstruction error. For this, in step 2, the examples are weighted for the
PCA by the responsibilities, and convergence is assessed by examining the change
in the log-likelihood of the data at each iteration. The soft version requires a choice
of O'z, the assumed variance in the directions orthogonal to the pancake.
The algorithm generates a set of local linear models for each digit. Given a test
pattern we evaluate the code length (the log-likelihood) against all the models for
all the digits. We use a hard method for classification - determining the identity of
the pattern only by the model which reconstructs it best. The absolute quality of the
best reconstruction and the relative qualities of slightly sub-optimal reconstructions
are available to reject ambiguous cases.
For a given linear model, not counting the code cost implies that deformations
of the images along the principal components for the sub-class are free. This is
like the metric used by [11] except that they explicitly specified the directions in
which deformations should be free, rather than learning them from the data. We
wished to incorporate information about the preferred directions without losing
the summarisation capacity of the local models, and therefore turned to the tangent
prop algorithm [12].
Tangent prop takes into account information about how the output of the system
should vary locally with particular distortions of the input by penalising the system
for having incorrect derivatives in the relevant directions. In our case, the overall
classification process is highly non-linear, making the application of tangent-prop
to it computationally unattractive, but it is easy to add a tangent constraint to the
reconstruction step because it is linear. Imagine requiring the system f(p) = A.p
to reconstruct x + .M and x -7-. t as well as x, for an input example x, and distortion
t (the tangent vector), and where 7-. is a weight. Their contribution to the error is
proportional to Ix - A.xl z + 7-. z lt - A.tI Z , where the second term is equivalent to
the error term that tangent-prop would add. Incorporating this into the PCA is as
simple as adding a weighted version of tt T to the covariance matrix - the tangent
vectors are never added to the means of the sub-classes.
3 Results
We have evaluated the performance of the system on data from the CEDAR
CDROM 1 database containing handwritten digits lifted from mail pieces passing through a United States Post Office [8]. We divided the br training set of binary
Recognizing Handwritten Digits Using Mixtures of Linear Models
1019
Table 1: Classification errors on the validation test when different weightings are
used for the tangent vectors during clustering and recognition. No rejections were
allowed.
segmented digits into 7,000 training examples, 2,000 validation examples and 2,000
"internal test" examples. All digits were equally represented in these sets. The
binary images in the database are all of different sizes, so they were scaled onto a
16 x 16 grid and then smoothed with a Gaussian filter.
The validation set was used to investigate different choices for the Gaussian filter
variances, the numbers of sub-classes per digit and principal components per
model, and the different weightings on the tangent vectors. Clearly this is a large
parameter space and we were only able to perform a very coarse search. In the
results reported here all digits have the same number of sub-classes (10) and the
number of principal components per model was picked so as to explain 95% of the
training set variance assigned to that model. Once a reasonable set of parameter
settings had been decided upon, we used all 11,000 images to train a final version
which was tested on the official bs set (2711 images).
There are two major steps to the algorithm; defining sub-classes within each digit
and reconstructing for recognition. We found that the tangent vectors should
be more heavily weighted in the sub-class clustering step than during ultimate
recognition. Figure 2 shows the means of the 10 sub-classes for the digit two where
the clustering has been done (a) without or (b) with tangent vectors. It is clear that
the clusters defined in (b) capture different styles of 2s in a way that those in (a)
do not - they are more diverse and less noisy. They also perform better. The raw
error rate (no rejections allowed) on the validation set with different amounts of
tangent vector weightings are shown in Table 1. The results on the official test set
(2711 examples) are shown in Table 2.
4 Discussion and Conclusions
A mixture of local linear models is an effective way to capture the underlying
styles of handwritten digits. The first few principal components (less than 20 on
256-dimensional data) extract a significant proportion of the variance in the images
within each sub-class. The resulting models classify surprisingly well without
requiring large amounts of either storage or computation during recognition. Further, it is computationally easy and demonstrably useful to incorporate tangent
information requiring reconstruction to be good for small transformations of the
sample patterns in a handful of pre-specified directions. Adding tangent information is exactly equivalent to replacing each digit by a Gaussian cloud of digits
perturbed along the tangent plane and is much more efficient than adding extra
Geoffrey E. Hinton, Michael Revow, Peter Dayan
1020
a)
b)
Figure 2: Cluster means for images of twos. (a) Without tangent vectors. (b) Cluster
means using translation, rotation and scaling tangent vectors.
stochastically perturbed examples. The weightings applied to reconstruction of
the tangent vectors are equivalent to the variances of the cloud.
There is an interesting relationship between mixtures of linear models and mixtures
of Gaussians. Fitting a mixture of full-covariance Gaussians is typically infeasible
when the dimensionality of the input image space, n, is high and the number of
training examples is limited. Each Gaussian requires n parameters to define its
mean and n(n + 1)/2 more parameters to define its covariance matrix. One way to
reduce the number of parameters is to use a diagonal covariance matrix but this is a
very poor approximation for images because neighbouring pixel intensities are very
highly correlated. An alternative way to simplify the covariance matrix is to flatten
the Gaussian into a pancake that has h dimensions within the pancake and n - h
dimensions orthogonal to the pancake. Within this orthogonal subspace we assume
an isotropic Gaussian distribution (ie a diagonal covariance matrix with equal,
small variances along the diagonal), so we eliminate (n - h)(n - h + 1)/2 degrees
of freedom which is nearly all the degrees of freedom of the full covariance matrix
if h is small compared with n. Not counting the mean, this leaves h(2n - h + 1) /2
degrees of freedom which is exactly the number of parameters required to define
the first h principal components. l Thus we can view PCA as a way of fiercely
constraining a full covariance Gaussian but nevertheless leaving it free to model
lThe h th principal component only requires n - h
orthogonal to all previous components.
+ 1 parameters because it must be
Recognizing Handwritten Digits Using Mixtures of Linear Models
?
?
..' I???. ?'? ?1??. ?. ?' '1 . . . ,
~it:I; ..f....
:~l ~
.???
. ..... ..... "
,,~:I!:::';;.?;
.I'"
.....?. ........
_._....??
:d...: .1.:
..........
f!~~,
-. :rL
..................
....o.?
1021
~
"'lii!;~"
?.
.
?
I;':~?~;
.,.
...
roo
r. ?
........ " .
..
........
a.
.-.
..... ., . ......
.o...
:H:fiiiiP.~!?
,..
.....
.. i;;::~?:
..::P.'a
..........
?.
..
'
..?? . .u
..
.?.
.
Li. a:!:?
.:t-ir:~:,.
.!si??. ~ .....??..
....'ii:
?~HiiiH~
a&,.
:-!'**!Hr:~:'
Figure 3: Reconstruction of a 2 (left column) using 2 (upper) and 0 (lower) models.
Linear 0 e
Linear Models + Tangent Dist
Tan ent Distance
Table 2: Classification errors on the official test test when no rejections are allowed.
Memory requirements are indicated in terms of the number of 16 x 16 images that
need to be stored.
important correlations.
We have investigated a number of extensions to the basic scheme. As only one of
these yielded improved results on the validation set, we will only briefly review
them. If instead of assuming an isotropic Gaussian within the pancake we use
an ellipsoidal subspace, then we can can take into account the different variances
along each of the h principal directions. This is akin to incorporating a code cost [14].
Similarly, the squared reconstruction error used in the basic scheme also assumes
an isotropic distribution. Again a diagonal covariance matrix may be substituted.
We surmise that this was not successful because we had insufficient training data
to estimate the variances along some dimensions reliably; for example some of the
edge pixels were never turned on in the training data for some models.
The Euclidean metric for the reconstruction error is convenient as it authorises the
use of a powerful methods such as principal components; however as pointed out
in the introduction it is deficient for situations like character recognition. We tried a
scheme in which the models were trained as described above, but tangent distance
[11] was used during testing. This method yielded marginally improved results
on both the validation and test sets (Table 2). More adventurous tangent options,
including using them in the clustering phase, were explored by [5, 10].
PCA models are not ideal as generative models of the data because they say nothing about how to generate values of components in the directions orthogonal
to the pancake. To ameliorate this, the generative model may be formulated as
f(p) = A.p + ?, where the components of ? are independent and the factors p have
some prior covariance matrix. The generative weights of this autoencoder can
1022
Geoffrey E. Hinton, Michael Revow, Peter Dayan
be obtained using the technique of maximum likelihood factor analysis (which is
closely related to PCA) and the resulting architecture and hierarchical variants of it
can be formulated as real valued versions of the Helmholtz machine [3,6]. The cost
of coding the factors relative to their prior is implicitly included in this formulation, as is the possibility that different input pixels are subject to different amounts
of noise. Unlike PCA, factor analysis privileges the particular input coordinates
(rather than being invariant to rotations of the input covariance matrix).
References
[1] Bourlard, H & Kamp, Y (1988). Auto-association by Multilayer Perceptrons and Singular
Value Decomposition. BioI. Cybernetics 59,291-294.
[2] Bregler, C & Omohundro, SM (1995). Non-linear image interpolation using surface
learning. This volume.
[3] Dayan, P, Hinton, GE, Neal, RM & Zemel, RS (1995). The Helmholtz machine. Neural
Computation, in press.
[4] Dempster, AP, Laird, NM & Rubin, DB (1976). Maximum likelihood from incomplete
data via the EM algorithm. Proceedings of the Royal Statistical Society, 1-38.
[5] Hastie, T, Simard, P & Sackinger, E (1995). Learning prototype models for tangent
distance. This volume.
[6] Hinton, GE, Dayan, P, Frey, BJ, Neal, RM (1995). The wake-sleep algorithm for unsupervised neural networks. Submitted for publication.
[7] Hinton, GE & Zemel, RS (1994). Autoencoders, minimum deSCription length and
Helmholtz free energy. In JD Cowan, G Tesauro & J Alspector, editors, Advances in
Neural Information Processing Systems 6. San Mateo, CA: Morgan Kaufmann.
[8] Hull, JJ (1994). A database for handwritten text recognition research. IEEE Transactions
on Pattern Analysis and Machine Intelligence, 16, 550-554.
[9] Kambhatla, N & Leen, TK (1994). Fast non-linear dimension reduction. In JD Cowan,
G Tesauro & J Alspector, editors, Advances in Neural Information Processing Systems 6.
San Mateo, CA: Morgan Kaufmann.
[10] Schwenk, H & Milgram, M (1995). Transformation invariant autoassociation with application to handwritten character recognition. This volume.
[11] Simard, P, Le Cun, Y & and Denker, J (1993). Efficient pattern recognition using a new
transformation distance. In SJ Hanson, JD Cowan & CL Giles, editors, Advances in
Neural Information Processing Systems 5, 50-58. San Mateo, CA: Morgan Kaufmann.
[12] Simard, P, Victorri, B, LeCun, Y & Denker, J (1992). Tangent Prop - A formalism for
specifying selected invariances in an adaptive network. In JE Moody, SJ Hanson & RP
Lippmann, editors, Advances in Neural Information Processing Systems 4. San Mateo, CA:
Morgan Kaufmann.
[13] Williams, CKI, Zemel, RS & Mozer, MC (1993). Unsupervised learning of object models.
In AAAI FalI1993 Symposium on Machine Learning in Computer Vision, 20-24.
[14] Zemel, RS (1993). A Minimum Description Length Framework for Unsupervised Learning.
PhD Dissertation, Computer Science, University of Toronto, Canada.
IThis research was funded by the Ontario Information Technology Research Centre and
NSERC. We thank Patrice Simard, Chris Williams, Rob Tibshirani and Yann Le Cun for
helpful discussions. Geoffrey Hinton is the Noranda Fellow of the Canadian Institute for
Advanced Research.
| 962 |@word briefly:1 version:9 compression:2 proportion:1 grey:1 r:4 tried:1 covariance:12 decomposition:1 reduction:1 initial:1 united:1 si:1 assigning:1 must:1 generative:8 leaf:1 intelligence:1 selected:1 plane:2 isotropic:4 summarisation:1 dissertation:1 provides:1 coarse:1 toronto:3 along:10 symposium:1 incorrect:1 fitting:2 expected:1 alspector:2 embody:2 dist:1 infelicitous:1 underlying:1 transformation:9 fellow:1 ti:1 exactly:2 scaled:1 rm:2 unit:1 local:10 frey:1 severely:1 iie:2 interpolation:1 ap:1 might:2 emphasis:1 mateo:4 specifying:1 autoassociation:1 limited:3 averaged:1 decided:1 responsible:1 lecun:1 testing:2 maximisation:1 digit:26 procedure:2 reject:1 convenient:1 pre:2 flatten:1 seeing:1 suggest:1 onto:2 close:1 storage:2 writing:3 equivalent:4 straightforward:2 williams:2 splitting:1 assigns:1 population:2 variation:4 coordinate:1 imagine:2 tan:1 heavily:1 losing:1 neighbouring:1 designing:1 helmholtz:3 recognition:12 surmise:1 database:3 cloud:2 capture:3 mozer:1 dempster:1 trained:1 segment:1 upon:1 schwenk:1 represented:2 train:3 fast:2 describe:1 effective:1 zemel:4 choosing:1 valued:1 distortion:2 roo:1 otherwise:1 reconstruct:1 encoder:1 say:1 itself:1 noisy:1 la4:1 final:1 laird:1 patrice:1 advantage:1 net:1 reconstruction:15 turned:2 relevant:1 ontario:2 description:5 ent:1 convergence:1 regularity:1 cluster:3 requirement:1 produce:1 tk:1 object:1 measured:3 nearest:2 wished:1 dividing:1 overnight:1 implies:1 differ:1 direction:11 thick:1 closely:1 filter:2 hull:1 assign:1 bregler:1 extension:1 great:1 bj:1 major:1 vary:1 kambhatla:1 estimation:1 label:1 weighted:3 clearly:1 gaussian:11 rather:7 lifted:1 office:1 publication:1 derived:1 properly:1 likelihood:6 helpful:1 dayan:7 entire:1 typically:2 eliminate:1 hidden:4 transformed:1 pixel:6 overall:2 among:1 classification:4 priori:1 equal:1 construct:1 never:2 having:2 once:1 unsupervised:3 nearly:1 summarise:1 others:1 develops:1 simplify:1 few:2 neighbour:2 cheaper:1 phase:1 privilege:1 linearised:1 freedom:3 highly:2 investigate:1 possibility:1 mixture:11 pc:1 behind:1 edge:1 encourage:1 orthogonal:5 pancake:8 incomplete:1 euclidean:3 deformation:3 classify:1 soft:2 column:1 giles:1 formalism:1 altering:1 assignment:1 cost:7 cedar:1 recognizing:4 examining:1 successful:1 reported:1 stored:1 encoders:3 perturbed:2 density:2 ie:1 retain:1 cki:1 michael:5 moody:1 squared:4 again:1 nm:1 aaai:1 reconstructs:2 choose:2 containing:1 worse:1 stochastically:1 creating:1 lii:1 simard:4 derivative:1 style:5 leading:1 return:1 li:1 account:2 coding:1 explicitly:1 piece:1 performed:3 view:1 picked:1 responsibility:3 option:1 contribution:1 ir:1 variance:9 kaufmann:4 kamp:1 misfit:1 handwritten:9 raw:1 marginally:1 mc:1 r4400:1 m5s:1 cybernetics:1 classified:1 submitted:1 za:1 explain:1 milgram:1 against:1 energy:1 stop:1 knowledge:1 penalising:1 improves:1 dimensionality:2 back:1 improved:3 formulation:1 evaluated:1 done:1 leen:1 just:6 autoencoders:2 correlation:1 ei:1 replacing:1 sackinger:1 quality:2 indicated:1 effect:2 requiring:3 assigned:2 neal:2 during:6 ambiguous:1 criterion:1 tt:1 omohundro:1 image:17 rotation:3 rl:1 volume:3 association:1 approximates:1 significant:1 grid:1 similarly:1 pointed:1 centre:1 had:2 funded:1 surface:1 add:2 own:1 tesauro:2 certain:1 binary:2 seen:1 minimum:4 morgan:4 somewhat:1 care:1 ii:1 full:3 reduces:1 smooth:1 segmented:1 cross:1 divided:1 post:1 equally:2 variant:2 basic:2 multilayer:1 vision:1 expectation:1 metric:4 iteration:1 represent:1 separately:1 victorri:1 wake:1 singular:1 leaving:1 appropriately:1 extra:1 unlike:1 subject:2 undergo:1 deficient:1 db:1 cowan:3 counting:2 ideal:1 constraining:1 canadian:1 enough:1 easy:2 fit:2 architecture:1 hastie:1 reduce:1 idea:3 prototype:1 br:1 minimise:1 motivated:1 pca:15 ultimate:1 akin:1 peter:5 passing:1 jj:1 reassign:1 dramatically:1 useful:1 clear:1 amount:4 ellipsoidal:1 locally:3 ten:1 demonstrably:2 generate:1 per:3 tibshirani:1 pace:1 diverse:1 didactic:1 four:2 nevertheless:1 sum:1 powerful:1 ameliorate:1 place:1 reasonable:1 yann:1 separation:1 recognise:1 scaling:1 bit:2 bound:1 sleep:1 yielded:2 constraint:3 handful:1 ri:1 generates:1 combination:1 poor:2 slightly:2 em:4 character:3 reconstructing:1 cun:2 rob:1 making:1 b:1 invariant:2 taken:1 computationally:3 turn:1 discus:1 ge:3 available:1 gaussians:3 apply:1 denker:2 hierarchical:1 appropriate:1 alternative:2 rp:1 jd:3 original:1 assumes:1 remaining:1 clustering:6 unfortunate:1 log2:1 deparbnent:1 build:1 approximating:1 society:1 added:1 usual:1 diagonal:4 amongst:1 subspace:6 distance:9 separate:1 thank:1 capacity:1 chris:1 mail:1 lthe:1 barely:1 reason:1 assuming:2 length:7 code:8 retained:1 relationship:1 insufficient:1 negative:1 rise:1 reliably:1 perform:4 upper:2 sm:1 defining:3 hinton:9 incorporated:1 situation:1 smoothed:1 canada:2 intensity:1 required:1 specified:3 hanson:2 able:1 pattern:16 cdrom:1 including:1 memory:1 royal:1 bourlard:1 hr:1 advanced:1 scheme:3 technology:1 picture:1 portray:1 autoencoder:12 auto:4 extract:1 text:2 prior:3 review:1 tangent:33 determining:1 relative:2 interesting:1 proportional:2 geoffrey:6 validation:7 degree:3 sufficient:1 rubin:1 editor:4 storing:1 translation:2 changed:1 surprisingly:1 free:7 infeasible:1 generalise:1 institute:1 absolute:1 distributed:1 curve:1 dimension:5 calculated:1 evaluating:1 collection:2 adaptive:1 projected:1 san:4 transaction:1 sj:2 lippmann:1 ignore:1 preferred:1 implicitly:1 generalising:1 assumed:1 discriminative:1 noranda:1 search:1 iterative:1 table:5 ca:4 investigated:1 cl:1 adventurous:1 official:3 substituted:1 whole:1 noise:1 nothing:2 allowed:4 je:1 sub:17 xl:1 weighting:4 ix:1 explored:1 unattractive:1 incorporating:4 quantization:1 recognising:1 adding:5 effectively:1 phd:1 illustrates:1 rejection:3 lt:1 nserc:1 corresponds:1 prop:6 bioi:1 viewed:2 identity:1 formulated:2 revow:5 change:2 hard:1 included:1 determined:1 except:1 ithis:1 principal:16 invariance:1 perceptrons:1 formally:1 internal:1 assessed:2 incorporate:2 evaluate:1 tested:1 gust:1 correlated:1 |
7,202 | 963 | An Auditory Localization
and Coordinate Transform Chip
Timothy K. Horiuchi
timmer@cns.caltech.edu
Computation and Neural Systems Program
California Institute of Technology
Pasadena, CA 91125
Abstract
The localization and orientation to various novel or interesting
events in the environment is a critical sensorimotor ability in all
animals, predator or prey. In mammals, the superior colliculus
(SC) plays a major role in this behavior, the deeper layers exhibiting topographically mapped responses to visual, auditory, and
somatosensory stimuli. Sensory information arriving from different modalities should then be represented in the same coordinate
frame. Auditory cues, in particular, are thought to be computed
in head-based coordinates which must then be transformed to retinal coordinates. In this paper, an analog VLSI implementation for
auditory localization in the azimuthal plane is described which extends the architecture proposed for the barn owl to a primate eye
movement system where further transformation is required. This
transformation is intended to model the projection in primates from
auditory cortical areas to the deeper layers of the primate superior
colliculus. This system is interfaced with an analog VLSI-based
saccadic eye movement system also being constructed in our laboratory.
Introduction
Auditory localization has been studied in many animals, particularly the barn owl.
Most birds have a resolution of only 10 to 20 degrees, but owls are able to orient
788
Timothy Horiuchi
to sound with an accuracy of 1 to 2 degrees which is comparable with humans.
One important cue for localizing sounds is the relative time of arrival of a sound
to two spatially separated ears. A neural architecture first described by Jeffress
(1948) for measuring this time difference has been shown to exist in the barn owl
auditory localization system (Konishi 1986) . An analog VLSI implementation of
the barn owl system constructed by Lazzaro (1990) is extended here to include a
transformation from head coordinates to retinal coordinates.
In comparison to the barn owl, the neurophysiology of auditory localization in cats
and primates is not as well understood and a clear map of auditory space does not
appear to be present in the inferior colliculus as it is in the owl. It has been suggested
that cortical auditory regions may provide the head-based map of auditory space
(Groh and Sparks 1992).
In primates, where much of the oculomotor system is based in retinotopic coordinates , head-based information must ultimately be transformed in order to be used.
While other models of coordinate transformation have been proposed for visual
information (e.g. Zipser and Andersen 1988, Krommenhoek et al. 1993) and for
auditory information (Groh and Sparks 1992), the model of coordinate transformation used in this system is a switching network which shifts the entire projection of
the head-based map of auditory space onto a retinotopic "colliculus" circuit. This
particular model is similar to a basis function approach where intermediate units
have compact receptive fields in an eye-position / head-based azimuth space and
the output units sum the outputs of a subset of these units.
The auditory localization system described here provides acoustic target information
to an analog VLSI-based saccadic eye movement system (Horiuchi, Bishofberger,
and Koch 1994) being developed in our laboratory for multimodal operation.
Bandpass
_::::::10-,.,.... . . . Filtering
@3.2kHz
Thresholded
Zero-crossings
electret
Localization
condenser
microphones
and
Bandpass
Filtering
@3.2kHz
Thresholded
Zero-crossings
Coordinate
~--Eye Position
Transform
Chip
Sound Source Location
(in retinal coordinates)
Figure 1: Block diagram of the auditory localization system. The analog front end
consists of external discrete analog electronics.
789
An Auditory Localization and Coordinate Transform Chip
M? Snop II 45 degrees R of Center
.'
..
.-
i
f
0
?1
..
.
"
~+-----~-----+----~~----+-----~----~
o
M? Snop II CerRr POIIition
i
I
?1
~+-----~-----+------~----+-----~----~
o
Finga' Snop 0145 degrees LofCCDtcr
I
f
..
'
: :
?1
~
: :~
..
. ..
". ~
~
:
:~
~~----~-----+----~~----+-----~----~
o
Figure 2: Filtered signals of the left and right microphones from 3 different angles.
790
Timothy Horiuchi
The Localization System
The analog front-end of the system (see figure 1) consists of three basic components,
the microphones, the filter stage, and the thresholded, zero-crossing stage. Two
microphones are placed with their centers about 2 inches apart. For any given
time difference in arrival of acoustic stimuli, there are many possible locations from
which the sound could have originated. These points describe a hyperbola with
the two microphones as the two foci. If the sound source is distant enough, we
can estimate the angle since the hyperbola approaches an asymptote. The current
system operates on a single frequency and the inter-microphone distance has been
chosen to be just under one wavelength apart at the filter frequency. The filter
frequency chosen was 3.2 kHz because the author's finger snap, used extensively
during development contained a large component at that frequency. The next step
in the computation consists of triggering a digital pulse at the moment of zerocrossing if the acoustic signal is large enough.
Thn::sb.olded Zct<Hlrossing Detection Pulses
6
7.5
5
5.0
25
4
'-- '--
~
.e
}
'-- '-- '-- '--
0.0
?25
~
?5.0
~
2
]
0
.?
.!!
1\
!<
it
'--
3
!.....
'"
..... I i
J'L
",... VV.....,
I!'
II
flA~
V V '"
V
?1
V
V
?75
?10.0
'R
II
~
?125
?2
-15.0
? 175
?3
0
2
3
4
5
6
1imo (aeconda) (10")
Figure 3: Example of output pulses from the external circuitry. Zero phase is chosen
to be the positive-slope zero-crossing. Top: Digital pulses are generated at the time
of zero phase for signals whose derivative is larger than a preset threshold. Bottom:
3.2 kHz Bandpass filtered signal for a finger snap.
Phase Detection and Coordinate Transform in Analog VLSI
The analog VLSI component of the system consists of two axon delay lines (Mead
1988) which propagate the left and right microphone pulse signals in opposing
directions in order to compute the cross correlation (see Fig 4.) The location of the
peak in this correlation technique represents the relative phase of the two signals.
An Auditory Localization and Coordinate Transform Chip
791
This technique is described in more detail and with more biological justification by
Lazzaro (1990). The current implementation contains 15 axon circuits in each delay
line. This is shown in figure 4. At each position in the correlation delay line are
logical AND circuits which output a logic one when there are two active axon units
at that location. Since these units only turn on for specific time delays, they define
auditory "receptive fields". The output of this subsystem are 15 digital lines which
are passed on to the coordinate transform.
Left
Channel
JlllL
Right
Channel
Jllll
Figure 4: Diagram of the double axon delay line which accepts digital spikes on the
inputs and propagates them across the array. Whenever two spikes meet, a pulse
is generated on the output AND units. The position of the AND circuit which gets
activated indicates the relative time of arrival of the left and right inputs. NOTE:
the actual circuit contains 15 axon units.
A
Head-based Auditory Units
left ear-....IIIUI=llI1l~..... - right ear
f3
'S
::>
s::
.g
? .-4
(I)
A
~
~
~
eye position
Retinotopic Auditory Units
Figure 5:
For the one-dimensional case described in this project, the appropriate transform
from head to retinal coordinates is a rotation which subtracts the eye position. The
eye position information on the chip is represented as a voltage which activates one
of the eye position units. The spatial pattern of activation from the auditory units
is then "steered" to the output stage with the appropriate shift. (See figure 5). This
792
Timothy Horiuchi
is similar to a shift scheme proposed by Pitts and McCulloch (1947) for obtaining
pitch invariance for chord recognition. The eye position units are constructed from
an array of "bump" circuits (Delbriick 1993) which compare the eye position voltage
with its local voltage reference. The two dimensional array of intermediate units
take the digital signal from the auditory units and switch the "bump" currents onto
the output lines. The output current lines drive the inputs of a centroid circuit.
The current implementation of the shift can be viewed as a basis function approach
where a population of intermediate units respond to limited "ball-like" regions in
the two-dimensional space of horizontal eye position and sound source azimuth
(head-coordinates). The output units then sum the outputs of only those intermediate units which represent the same retinal location. It should be noted that this
coordinate transformation is closely related to the "dendrite model" proposed for
the projection of cortical auditory information to the deep SC by Groh and Sparks
(1992).
The final output stage converts this spatial array of current carrying lines into
a single output voltage which represents the centroid of the stimulus in retinal
coordinates. This centroid circuit (DeWeerth 1991) is intended to represent the
primate SC where a similar computation is believed to occur.
Results and Conclusions
Figure 6 shows three plots of the chip's output voltage as a function of the interpulse time interval. Figure 7 shows three plots of the full system's output voltage for
different eye position voltages. The output is roughly linear with azimuth and linear
with eye position voltage. In operation, the system input consists of a sound entering
the two microphones and the output consists of an analog voltage representing the
position of the sound source and a digital signal indicating that the analog data is
valid.
The auditory localization system described here is currently in use with an analog
VLSI-based model of the primate saccadic system to expand its operation into
the auditory domain (Horiuchi, Bishofberger, & Koch 1994). In addition to the
effort of our laboratory to model and understand biological computing structures
in real-time systems, we are exploring the use of these low power integrated sensors
in portable applications such as mobile robotics. Analog VLSI provides a compact
and efficient implementation for many neuromorphic computing architectures which
can potentially be used to provide, small, fast, low power sensors for a wide variety
of applications.
Acknowledgements
The author would like to acknowledge Prof. Christof Koch for his academic support
and use of laboratory facilities for this project, Brooks Bishofberger for his assistance
in constructing some of the discrete electronics and Prof. Carver Mead for running
the CNS184 course under which this chip was fabricated.
The author is supported by an AASERT grant from the Office of Naval Research.
793
An Auditory Localization and Coordinate Transform Chip
Output vs. Arrival Time Difference (3 eye positions)
3.1
3.0
2.9
6'
..;
old
2.8
..j
N
d
N
2.7
II
0
?
2.6
0
1
0
2.5
>
<;;
g
2.4
2.3
2.2
2.1
-600
-200
o
200
400
600
R to L delay (microseconda)
Figure 6: Chip output vs. input pulse timing: The chip was driven with a signal
generator and the output voltage was plotted for three different eye position voltages. Due to the discretized nature of the axon, there are only 15 axon locations at
which pulses can meet. This creates the staircase response.
References
T. Delbriick (1993) Investigations of Analog VLSI Visual Transduction and Motion
Processing, Ph.D. Thesis, California Institute of Technology
J. Groh and D. Sparks (1992) 2 Models for Transforming Auditory Signals from
Head-Centered to Eye-Centered Coordinates Bioi. Cybern. 67(4) 291-302.
T. Horiuchi, B. Bishofberger, & C. Koch, (1994) An Analog VLSI-based Saccadic
System, In (ed.), Advances in Neural Information Processing Systems 6 San Mateo,
CA: Morgan Kaufman
L. A. Jeffress (1948) A Place Theory of Sound Localization J. Comp_ Physiol.
Psychol. 41: 35-39.
M. Konishi (1986) Centrally Synthesized Maps of Sensory Space. TINS April, pp.
163-168.
K. P. Krommenhoek, A. J. Van Opstal, C. C. A. M. Gielen, J. A. ,M. Van Gisbergen.
(1993) Remapping of Neural Activity in the Motor Colliculus: A Neural Network
Study. Vision Research 33(9):1287-1298.
Timothy Horiuchi
794
J. Lazzaro. (1990) Silicon Models of Early Audition, Ph.D. Thesis, California Institute of Technology
C. Mead, (1988) Analog VLSI and Neural Systems Menlo Park: Addison-Wesley
W. Pitts and W. S. McCulloch, (1947) How we know universals: the perception of
auditory and visual forms. Bulletin of Mathematical Biophysics 9:127-147.
D. Zipser and R. A. Andersen (1988) A back-propagation programmed network
that simulates response properties of a subset of posterior parietal neurons. Nature
331:679-684.
Localization Output vs. Sound Source Azimuth
3.6
3.4
3.2
3.0
II
-e,.
on
..;
28
'Ij
N
'Ij
II
t
S
-e
26
24
22
~
20
1.8
1.6
1.4
?20
0
Figure 7: Performance of the full system on continuous input (sinusoidal) delivered
by a speaker from different angles. Note that 90 degrees denotes the center position.
The three plots are the outputs for three different settings of the eye position input
voltage.
| 963 |@word neurophysiology:1 azimuthal:1 propagate:1 pulse:8 mammal:1 moment:1 electronics:2 contains:2 olded:1 current:6 activation:1 must:2 physiol:1 distant:1 motor:1 asymptote:1 plot:3 v:3 cue:2 plane:1 filtered:2 provides:2 location:6 mathematical:1 constructed:3 consists:6 inter:1 roughly:1 behavior:1 discretized:1 actual:1 project:2 retinotopic:3 circuit:8 remapping:1 mcculloch:2 kaufman:1 developed:1 transformation:6 fabricated:1 unit:17 grant:1 appear:1 christof:1 timmer:1 positive:1 understood:1 local:1 timing:1 switching:1 mead:3 meet:2 bird:1 studied:1 mateo:1 limited:1 programmed:1 block:1 area:1 universal:1 thought:1 projection:3 get:1 onto:2 subsystem:1 cybern:1 map:4 center:3 resolution:1 spark:4 array:4 his:2 konishi:2 population:1 coordinate:21 justification:1 target:1 play:1 crossing:4 recognition:1 particularly:1 bottom:1 role:1 region:2 movement:3 chord:1 environment:1 transforming:1 ultimately:1 carrying:1 topographically:1 localization:16 creates:1 basis:2 multimodal:1 chip:10 various:1 represented:2 cat:1 finger:2 separated:1 horiuchi:8 fast:1 describe:1 jeffress:2 sc:3 whose:1 larger:1 snap:2 aasert:1 ability:1 transform:8 final:1 delivered:1 double:1 groh:4 ij:2 somatosensory:1 exhibiting:1 direction:1 closely:1 filter:3 centered:2 human:1 owl:7 investigation:1 biological:2 exploring:1 koch:4 barn:5 pitt:2 circuitry:1 bump:2 major:1 early:1 currently:1 activates:1 sensor:2 mobile:1 voltage:12 office:1 focus:1 naval:1 indicates:1 centroid:3 sb:1 entire:1 integrated:1 pasadena:1 vlsi:11 expand:1 transformed:2 zerocrossing:1 orientation:1 development:1 animal:2 spatial:2 field:2 f3:1 represents:2 park:1 stimulus:3 intended:2 phase:4 cns:1 opposing:1 detection:2 activated:1 carver:1 old:1 plotted:1 localizing:1 measuring:1 neuromorphic:1 subset:2 delay:6 azimuth:4 front:2 peak:1 thn:1 andersen:2 thesis:2 ear:3 external:2 steered:1 derivative:1 audition:1 sinusoidal:1 retinal:6 opstal:1 predator:1 slope:1 accuracy:1 interfaced:1 inch:1 hyperbola:2 drive:1 whenever:1 ed:1 sensorimotor:1 frequency:4 pp:1 auditory:28 logical:1 back:1 wesley:1 response:3 april:1 just:1 stage:4 correlation:3 deweerth:1 horizontal:1 propagation:1 fla:1 staircase:1 facility:1 spatially:1 entering:1 laboratory:4 assistance:1 during:1 inferior:1 noted:1 speaker:1 motion:1 novel:1 superior:2 rotation:1 khz:4 analog:16 synthesized:1 silicon:1 posterior:1 apart:2 driven:1 caltech:1 morgan:1 signal:10 ii:7 full:2 sound:11 academic:1 cross:1 believed:1 biophysics:1 pitch:1 basic:1 vision:1 represent:2 robotics:1 addition:1 interval:1 diagram:2 source:5 modality:1 simulates:1 zipser:2 intermediate:4 enough:2 switch:1 variety:1 architecture:3 triggering:1 shift:4 bishofberger:4 passed:1 effort:1 lazzaro:3 deep:1 clear:1 extensively:1 ph:2 exist:1 discrete:2 threshold:1 prey:1 thresholded:3 sum:2 colliculus:5 convert:1 orient:1 angle:3 respond:1 extends:1 place:1 comparable:1 layer:2 centrally:1 activity:1 occur:1 ball:1 across:1 primate:7 turn:1 addison:1 know:1 end:2 operation:3 appropriate:2 top:1 running:1 include:1 denotes:1 prof:2 spike:2 receptive:2 saccadic:4 distance:1 mapped:1 portable:1 potentially:1 implementation:5 gisbergen:1 neuron:1 acknowledge:1 parietal:1 extended:1 head:10 frame:1 delbriick:2 required:1 california:3 acoustic:3 accepts:1 brook:1 able:1 suggested:1 perception:1 pattern:1 oculomotor:1 program:1 power:2 event:1 critical:1 representing:1 scheme:1 technology:3 eye:18 psychol:1 acknowledgement:1 relative:3 interesting:1 filtering:2 generator:1 digital:6 degree:5 propagates:1 course:1 placed:1 supported:1 arriving:1 deeper:2 vv:1 institute:3 understand:1 wide:1 bulletin:1 van:2 cortical:3 valid:1 sensory:2 author:3 san:1 subtracts:1 compact:2 logic:1 active:1 continuous:1 channel:2 nature:2 ca:2 menlo:1 obtaining:1 dendrite:1 constructing:1 domain:1 arrival:4 fig:1 transduction:1 axon:7 position:18 originated:1 bandpass:3 tin:1 specific:1 timothy:5 wavelength:1 gielen:1 visual:4 contained:1 bioi:1 zct:1 viewed:1 operates:1 preset:1 microphone:8 invariance:1 indicating:1 support:1 |
7,203 | 964 | An Input Output HMM Architecture
Yoshua Bengio*
Dept. Informatique et Recherche
Operationnelle
Universite de Montreal, Qc H3C-3J7
bengioyOIRO.UMontreal.CA
Paolo Frasconi
Dipartimento di Sistemi e Informatica
Universita di Firenze (Italy)
paoloOmcculloch.ing.unifi.it
Abstract
We introduce a recurrent architecture having a modular structure
and we formulate a training procedure based on the EM algorithm.
The resulting model has similarities to hidden Markov models, but
supports recurrent networks processing style and allows to exploit
the supervised learning paradigm while using maximum likelihood
estimation.
1
INTRODUCTION
Learning problems involving sequentially structured data cannot be effectively dealt
with static models such as feedforward networks. Recurrent networks allow to model
complex dynamical systems and can store and retrieve contextual information in
a flexible way. Up until the present time, research efforts of supervised learning
for recurrent networks have almost exclusively focused on error minimization by
gradient descent methods. Although effective for learning short term memories,
practical difficulties have been reported in training recurrent neural networks to
perform tasks in which the temporal contingencies present in the input/output
sequences span long intervals (Bengio et al., 1994; Mozer, 1992).
Previous work on alternative training algorithms (Bengio et al., 1994) could suggest
that the root of the problem lies in the essentially discrete nature of the process
of storing information for an indefinite amount of time. Thus, a potential solution
is to propagate, backward in time, targets in a discrete state space rather than
differential error information. Extending previous work (Bengio & Frasconi, 1994a),
in this paper we propose a statistical approach to target propagation, based on the
EM algorithm. We consider a parametric dynamical system with discrete states and
we introduce a modular architecture, with subnetworks associated to discrete states.
The architecture can be interpreted as a statistical model and can be trained by the
EM or generalized EM (GEM) algorithms (Dempster et al., 1977), considering the
internal state trajectories as missing data. In this way learning is decoupled into
? also, AT&T Bell Labs, Holmdel, N J 07733
428
Yoshua Bengio, Paolo Frasconi
a temporal credit assignment subproblem and a static learning subproblem that
consists of fitting parameters to the next-state and output mappings defined by the
estimated trajectories . In order to iteratively tune parameters with the EM or GEM
algorithms, the system propagates forward and backward a discrete distribution over
the n states, resulting in a procedure similar to the Baum-Welch algorithm used
to train standard hidden Markov models (HMMs) (Levinson et al., 1983) . HMMs
however adjust their parameters using unsupervised learning, whereas we use EM
in a supervised fashion . Furthermore, the model presented here could be called
Input/Output HMM, or IOHMM , because it can be used to learn to map input
sequences to output sequences (unlike standard HMMs, which learn the output
sequence distribution) . This model can also be seen as a recurrent version of the
Mixture of Experts architecture (Jacobs et al. , 1991) , related to the model already
proposed in (Cacciatore and Nowlan, 1994). Experiments on artificial tasks (Bengio
& Frasconi , 1994a) have shown that EM recurrent learning can deal with long
term dependencies more effectively than backpropa~ation through time and other
alternative algorithms. However, the model used in (Bengio & Frasconi, 1994a) has
very limited representational capabilities and can only map an input sequence to a
final discrete state. In the present paper we describe an extended architecture that
allows to fully exploit both input and output portions of the data, as required by
the supervised learning paradigm. In this way, general sequence processing tasks,
such as production, classification, or prediction, can be dealt with.
2
THE PROPOSED ARCHITECTURE
We consider a discrete state dynamical system based on the following state space
x - f(x
U )
description:
t t-l, t
(1)
Yt = 9(Xt, Ut)
where Ut E R m is the input vector at time t, Yt E R r is the output vector, and
Xt E {I , 2, .. . , n} is a discrete state. These equations define a generalized Mealy
finite state machine, in which inputs and outputs may take on continuous values . In
this paper , we consider a probabilistic version of these dynamics, where the current
inputs and the current state distribution are used to estimate the state distribution
and the output distribution for the next time step. Admissible state transitions will
be specified by a directed graph 9 whose vertices correspond to the model 's states
and the set of successors for state j is Sj .
The system defined by equations (1) can be modeled by the recurrent architecture
depicted in Figure l(a) . The architecture is composed by a set of state n etworks
N j, j = 1 .. . n and a set of output networks OJ, j = 1 . .. n. Each one of the state
and output networks is uniquely associated to one of the states,and all networks
share the same input Ut . Each state network M has the task of predicting the next
state distribution , based on the current input and given that Xt-l = j . Similarly,
each output network OJ predicts the output of the system, given the current state
and input. All the subnetworks are assumed to be static and they are defined by
means of smooth mappings Nj (Ut ;9j ) and OJ (Ut; 1Jj), where 9j and 1Jj are vectors
of adjustable parameters (e .g., connection weights). The ranges of the functions
N j 0 may be constrained in order to account for the underlying transition graph
9 . Each output 'Pij ,t of the state subnetwork Nj (at time t) is associated to one
of the successors i of state j . Thus the last layer of M has as many units as the
cardinality of Sj. For convenience of notation, we suppose that 'Pij,t are defined for
each i, j = 1, ... , n and we impose the condition 'Pij ,t = 0 for each i not belonging
to S j . The softmax function is used in the last layer: 'Pij ,t = e a,j ,t ILlEsj ea l j,t, j =
1, ... , n , i E Sj where aij, t are intermediate variables that can be thought of as the
An Input Output HMM Architecture
cu ... nt Input
EIYt
...
Xt-l
current ??pectod output,
given PIlat Input Mquenc.
11 t
429
current atilt. dlatrtbutton
Ct= Pl' t I Ul )
lull
...
...
Xt
Xt+l
1 1 1
Yt-l
Yt
Yt+l
Xt-l
X(
Xt+l
'-'t
\.Yt-l
\Yt
I
I
Ut-l
HMM
\.Yt+l
I
Ut
.. ....
IOHMM
Ut+l
(b)
(a)
Figure 1: (a): The proposed IOHMM architecture. (b): Bottom: Bayesian network
expressing conditional dependencies for an IOHMM; top: Bayesian network for a
standard HMM
't
activations of the output units of subnetwork N j . In this way L:7=1 'Pij ,t = 1 Tij,t.
The vector
E R n represents the internal state of the model and it is computed as
a linear combination of the outputs of the state networks, gated by the previously
n
computed internal state:
't
= L ( j ,t-IIPj,t
(2)
j=1
where IPj,t = ['PIj,t, ... , 'Pnj,t]'. Output networks compete to predict the global
output of the system 1Jt E R r :
n
1Jt
= L (jt1Jjt
(3)
j=1
where 1Jjt E R r is the output of subnetwork OJ. At this level , we do not need
to further specify the internal architecture of the state and output subnetworks.
Depending on the task, the designer may decide whether to include hidden layers
and what activation rule to use for the hidden units.
This connectionist architecture can be also interpreted as a probability model. Let
us assume a multinomial distribution for the state variable Xt and let us consider
the main variable of the temporal recurrence (2). If we initialize the vector
to positive numbers summing to 1, it can be interpreted as a vector of initial state
probabilities. In general, we obtain relation (it = P(Xt = i I
having denoted
with ui the subsequence of inputs from time 1 to t, inclusively. Equation (2) then
has the following probabilistic interpretation:
't,
un,
'0
n
P(Xt
= i
lui) =
L
P(Xt
= i I Xt-I = j, ut}P(Xt-1 = j
lui-I)
(4)
j=l
i.e., the subnetworks N j compute transition probabilities conditioned on the input
sequence Ut:
P( Xt = ~. I Xt-l = ),.Ut
)
(5)
'Pij,t =
As in neural networks trained to minimize the output squared error, the output
1Jt of this architecture can be interpreted as an expected "position parameter"
for the probability distribution of the output Yt. However, in addition to being
conditional on an input Ut, this expectation is also conditional on the state Xt, i.e.
Yoshua Bengio, Paolo Frasconi
430
= E[Yt I Xt,Ut] . The actual form of the output density, denoted !Y(Yt ; 7]t), will
be chosen according to the task . For example a multinomial distribution is suitable
for sequence classification, or for symbolic mutually exclusive outputs. Instead, a
Gaussian distribution is adequate for producing continuous outputs. In the first
case we use a softmax function at the output of subnetworks OJ; in the second case
we use linear output units for the subnetworks OJ.
7]t
In order to reduce the amount of computation, we introduce an independency model
among the variables involved in the probabilistic interpretation of the architecture.
We shall use a Bayesian network to characterize the probabilistic dependencies
among these variables. Specifically, we suppose that the directed acyclic graph
9 depicted at the bottom of Figure 1b is a Bayesian network for the dependency
One of the most evident consequences
model associated to the variables u
of this independency model is that only the previous state and the current input are
relevant to determine the next-state. This one-step memory property is analogue
to the Markov assumption in hidden Markov models (HMM). In fact, the Bayesian
network for HMMs can be obtained by simply removing the Ut nodes and arcs from
them (see top of Figure Ib) .
I ,xI, YI.
3
A SUPERVISED LEARNING ALGORITHM
The learning algorithm for the proposed architecture is derived from the maximum
likelihood principle. The training data are a set of P pairs of input/ output sequences
(of length Tp): 1) = {(uip(p),Yip(p));p = 1 .. . P}. Let ?J denote the vector of
parameters obtained by collecting all the parameters (Jj and iJi of the architecture.
The likelihood function is then given by
p
L(?J; 1))
= II p(Yip(p) I uip(p); ?J).
(6)
p=l
The output values (used here as targets) may also be specified intermittently. For
example, in sequence classification tasks , one may only be interested in the output YT at the end of each sequence. The modification of the likelihood to account
for intermittent targets is straightforward. According to the maximum likelihood
principle, the optimal parameters are obtained by maximizing (6). In order to
apply EM to our case we begin by noting that the state variables Xt are not observed. Knowledge of the model's state trajectories would allow one to decompose
the temporal learning problem into 2n static learning subproblems. Indeed , if Xt
were known, the probabilities (it would be either 0 or 1 and it would be possible
to train each subnetwork separately, without taking into account any temporal dependency. This observation allows to link EM learning to the target propagation
approach discussed in the introduction . Note that if we used a Viterbi-like approximation (i .e., considering only the most likely path) , we would indeed have 2n static
learning problems at each epoch. In order to we derive the learning equations, let
us define the complete data as 1)c = {(uiP(p),yiP(p),xiP(p));p = 1 ... P}. The
corresponding complete data l%-likelihood is
T
T
T
Ic(?J;1)c) = """
L...IOgP(YIP(P),ZlP(P)
I u1P(p);
?J).
(7)
p=l
Since lc( ?J; 1)c) depends on the hidden state variables it cannot be maximized directly. The MLE optimization is then solved by introducing the auxiliary function
Q(?J; 0) and iterating the following two,steps for k = 1, 2 ... :,
Estimation:
Compute Q(?J ; ?J) = E[lc(?J; 1)c) r 1), ?J]
(8)
Maximization: Update the parameters as 0 t- arg max?J Q( ?J; 0)
An Input Output HMM Architecture
431
The expectation of (7) can be expressed as
p
Q(0 ; 0)
Tp
N
N
= L: L: L: (it!og P(Yt I Xt = i , Ut i 0) + L: hij,tlog<Pij ,t
p=1 t=1 i=1
(9)
j=1
I
where hij ,t = EIZitzj,t-l uf, yf; 0J, denoting Zit for an indicator variable = 1 if
Xt = i and 0 otherwise. The hat in ( it and hij ,t means that these variables are
computed using the "old" parameters 0 . In order to compute hij ,t we introduce
the forward probabilities Qit = P(YL Xt = i ; uD and the backward probabilities
f3it = p(yf I Xt = i ,
that are updated as follows:
un,
f3it = fY(Yt;l1it) Lt <Pti(Ut+df3l,t+l
(10)
Qit = fY(Yt; l1it) Lt <pa(ut} Qt ,t-l .
(11)
- f3it Qj ,t-l<Pij (ut)
"
wi QiT
Each iteration of the EM algorithm requires to maximize Q(0 ; 0). We first
consider a simplified case, in which the inputs are quantized (i .e., belonging
to a finite alphabet {0"1, "" O"K}) and the subnetworks behave like lookup tables addressed by the input symbols O"t, i.e. we interpret each parameter as
Wi' k = P(Xt = i I Xt-l = j , O"t = k). For simplicity, we restrict the analysis to classification tasks and we suppose that targets are specified as desired final states for
each sequence. Furthermore, no output subnetworks are used in this particular
application of the algorithm. In this case we obtain the reestimation formulae:
h
Wijk =
ij ,t -
"
"p"
(12)
. .
{j,t(j,t-l
w i ESj wp=1 wt :Ut=k , + T
?
"'T '
In general, however, if the subnetworks have hidden sigmoidal units, or use a softmax function to constrain their outputs to sum to one, the maximum of Q cannot
be found analytically. In these cases we can resort to a GEM algorithm, that simply produces an increase in Q, for example by gradient ascent. In this case, the
derivatives of Q with respect to the parameters can be easily computed as follows.
Let Ojlt be a generic weight in the state subnetwork N j . From equation (9):
8Q(0;0)
80jk
= L:L:L:hij,t_l_8<pij ,t
p
t
i
<Pij,t 80jk
(13)
where the partial derivatives &:e~;t can be computed using backpropagation . Similarly, denoting with t'Jik a generic weight of the output subnetwork Oi, we have:
8Q( 0; 0)
'" '" '" .
8
87]a t
8t'J ?
= L..JL..JL..J(i , t~logfY(Yy;l1it) 8t'J .'
p
t
t
7],t,t
,k
,k
(14)
where ~;;:~t are also computed using backpropagation. Intuitively, the parameters
are updated as if the estimation step of EM had provided targets for the outputs of
the 2n subnetworks , for each time t . Although GEM algorithms are also guaranteed
to find a local maximum of the likelihood, their convergence may be significantly
slower compared to EM. In several experiments we noticed that convergence can be
accelerated with stochastic gradient ascent .
432
4
Yoshua Bengio, Paolo Frasconi
COMPARISONS
It appears natural to find similarities between the recurrent architecture described
so far and standard HMMs (Levinson et al., 1983). The architecture proposed in this
paper differs from standard HMMs in two respects: computing style and learning.
With IOHMMs, sequences are processed similarly to recurrent networks, e.g., an
input sequence can be synchronously transformed into an output sequence. This
computing style is real-time and predictions of the outputs are available as the input
sequence is being processed. This architecture thus allows one to implement all three
fundamental sequence processing tasks: production, prediction, and classification.
Finally, transition probabilities in standard HMMs are fixed, i.e. states form a
homogeneous Markov chain. In IOHMMs, transition probabilities are conditional
on the input and thus depend on time, resulting in an inhomogeneous Markov chain.
Consequently, the dynamics of the system (specified by the transition probabilities)
are not fixed but are adapted in time depending on the input sequence.
The other fundamental difference is in the learning procedure. While interesting
for their capabilities of modeling sequential phenomena, a major weakness of standard HMMs is their poor discrimination power due to unsupervised learning. An
approach that has been found useful to improve discrimination in HMMs is based
on maximum mutual information (MMI) training . It has been pointed out that
supervised learning and discriminant learning criteria like MMI are actually strictly
related (Bridle, 1989). Although the parameter adjusting procedure we have defined
is based on MLE,
is used as desired output in response to the input
resulting
in discriminant supervised learning. Finally, it is worth mentioning that a number
of hybrid approaches have been proposed to integrate connectionist approaches into
the HMM frame\'Vork. For example in (Bengio et al. , 1992) the observations used
by the HMM are generated by a feedforward neural network. In (Bourlard and
Wellekens, 1990) a feedforward network is used to estimate state probabilities, conditional to the acoustic sequence. A common feature of these algorithms and the
one proposed in this paper is that neural networks are used to extract temporally
local information whereas a Markovian system integrates long-term constraints.
yf
uf ,
We can also establish a link between IOHMMs and adaptive mixtures of experts
(ME) (Jacobs et al., 1991). Recently, Cacciatore & Nowlan (1994) have proposed a
recurrent extension to the ME architecture, called mixture of controllers (MC), in
which the gating network has feedback connections, thus allowing to take temporal
context into account . Our IOHMM architecture can be interpreted as a special case
of the MC architecture, in which the set of state subnetworks play the role of a
gating network having a modular structure and second order connections.
5
REGULAR GRAMMAR INFERENCE
In this section we describe an application of our architecture to the problem of
grammatical inference. In this task the learner is presented a set of labeled strings
and is requested to infer a set of rules that define a formal language. It can be
considered as a prototype for more complex language processing problems. However,
even in the "simplest" case, i.e. regular grammars, the task can be proved to
be NP-complete (Angluin and Smith, 1983) . We report experimental results on
a set of regular grammars introduced by Tomita (1982) and afterwards used by
other researchers to measure the accuracy of inference methods based on recurrent
networks (Giles et al. , 1992; Pollack, 1991; Watrous and Kuhn , 1992) .
We used a scalar output with supervision on the final output YT that was modeled
as a Bernoulli variable fy (YT ; 7]T) = 7]~T (1 - 7] ) l-YT, with YT = 0 if the string
is rejected and YT = 1 if it is accepted . In tbis application we did not apply
An Input Output HMM Architecture
433
Table 1: Summary of experimental results on the seven Tomita's grammars.
Grammar
n*
1
2
3
4
5
6
7
2
8
7
4
4
3
3
Sizes
FSA min
2
3
5
4
4
3
5
Convergence
.600
.800
.150
.100
.100
.350
.450
Average
1.000
.965
.867
1.000
1.000
1.000
.856
Accuracies
Worst
Best
1.000 1.000
.834 1.000
.775 1.000
1.000 1.000
1.000 1.000
1.000 1.000
.815 1.000
W&K Best
1.000
1.000
.783
.609
.668
.462
.557
external inputs to the output networks . This corresponds to modeling a Moore
finite state machine. Given the absence of prior knowledge about plausible state
paths, we used an ergodic transition graph (i.e., fully connected).In the experiments
we measured convergence and generalization performance using different sizes for
the recurrent architecture. For each setting we ran 20 trials with different seeds
for the initial weights. We considered a trial successful if the trained network was
able to correctly label all the training strings. The model size was chosen using a
cross-validation criterion based on performance on 20 randomly generated strings
of length T ::; 12. For comparison, in Table 1 we also report for each grammar
the number of states of the minimal recognizing FSA (Tomita, 1982). We tested
the trained networks on a corpus of 213 - 1 binary strings of length T ::; 12. The
final results are summarized in Table 1. The column "Convergence" reports the
fraction of trials that succeeded to separate the training set. The next three columns
report averages and order statistics (worst and best trial) of the fraction of correctly
classified strings, measured on the successful trials. For each grammar these results
refer to the model size n* selected by cross-validation. Generalization was always
perfect on grammars 1,4,5 and 6. For each grammar, the best trial also attained
perfect generalization. These results compare very favorably to those obtained with
second-order networks trained by gradient descent, when using the learning sets
proposed by Tomita. For comparison , in the last column of Table 1 we reproduce
the results reported by Watrous & Kuhn (1992) in the best of five trials. In most
of the successful trials the model learned an actual FSA behavior with transition
probabilities asymptotically converging either to 0 or to 1. This renders trivial the
extraction of the corresponding FSA . Indeed, for grammars 1,4,5, and 6, we found
that the trained networks behave exactly like the minimal recognizing FSA .
A potential training problem is the presence of local maxima in the likelihood function. For example, the number of converged trials for grammars 3, 4, and 5 is quite
small and the difficulty of discovering the optimal solution might become a serious
restriction for tasks involving a large number of states. In other experiments (Bengio & Frasconi , 1994a) , we noticed that restricting the connectivity of the transition
graph can significantly help to remove problems of convergence. Of course, this approach can be effectively exploited only if some prior knowledge about the state
space is available. For example, applications of HMMs to speech recognition always
rely on structured topologies.
6
CONCLUSIONS
There are still a number of open questions . In particular, the effectiveness of the
model on tasks involving large or very large state spaces needs to be carefully evaluated. In (Bengio & Frasconi 1994b) we show that learning long term dependencies
in these models becomes more difficult as we increase the connectivity of the state
434
Yoshua Bengio, Paolo Frasconi
transition graph. However, because transition probabilities of IOHMMs change at
each t, they deal better with this problem of long-term dependencies than standard
HMMs. Another interesting aspect to be investigated is the capability of the model
to successfully perform tasks of sequence production or prediction. For example,
interesting tasks that could also be approached are those related to time series
modeling and motor control learning.
References
Angluin, D. and Smith, C. (1983). Inductive inference: Theory and methods. Computing Surveys, 15(3):237-269.
Bengio, Y. and Frasconi, P. (1994a) . Credit assignment through time: Alternatives
to backpropagation. In Cowan, J., Tesauro, G., and Alspector, J., editors,
Advances in Neural Information Processing Systems 6. Morgan Kaufmann.
Bengio, Y. and Frasconi, P. (1994b). An EM Approach to Learning Sequential
Behavior. Tech. Rep. RT-DSI/11-94, University of Florence.
Bengio, Y., De Mori, R., Flammia, G., and Kompe, R. (1992). Global optimization
of a neural network-hidden markov model hybrid. IEEE Transactions on Neural
Networks, 3(2):252-259.
Bengio, Y., Simard, P., and Frasconi, P. (1994) . Learning long-term dependencies
with gradient descent is difficult. IEEE Trans. Neural Networks, 5(2).
Bourlard, H. and Wellekens, C. (1990). Links between hidden markov models and
multilayer perceptrons. IEEE Trans. Pattern An. Mach. Intell., 12:1167-1178.
Bridle, J. S. (1989). Training stochastic model recognition algorithms as networks can lead to maximum mutual information estimation of parameters. In
D.S.Touretzky, ed., NIPS2, pages 211-217. Morgan Kaufmann.
Cacciatore, T. W. and Nowlan, S. J. (1994). Mixtures of controllers for jump
linear and non-linear plants. In Cowan, J. et. al., editors, Advances in Neural
Information Processing Systems 6, San Mateo, CA. Morgan Kaufmann.
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum-likelihood from
incomplete data via the EM algorithm. J. Royal Stat. Soc. B,39:1-38.
Giles, C. L., Miller, C. B ., Chen, D., Sun, G. Z., Chen, H. H., and Lee, Y. C. (1992).
Learning and extracting finite state automata with second-order recurrent neural networks. Neural Computation, 4(3):393-405.
Jacobs, R. A., Jordan, M. I., Nowlan, S. J., and Hinton, G. E. (1991). Adaptive
mixture of local experts. Neural Computation, 3:79-87.
Levinson, S. E., Rabiner, L. R., and Sondhi, M. M. (1983). An introduction to
the application of the theory of probabilistic functIons of a markov process to
automatic speech recognition. Bell System Technical Journal, 64(4):1035-1074.
Mozer, M. C. (1992). The induction of multiscale temporal structure. In Moody,
J. et. al., eds, NIPS 4 pages 275-282. Morgan Kaufmann.
Pollack, J. B. (1991) . The induction of dynamical recognizers. Machine Learning,
7(2):196-227.
Tomita, M. (1982). Dynamic construction of finite-state automata from examples
using hill-climbing. Proc. 4th Cog. Science Con!, pp. 105-108, Ann Arbor MI.
Watrous, R. 1. and Kuhn, G. M. (1992). Induction of finite-state languages using
second-order recurrent networks. Neural Computation, 4(3):406-414 .
| 964 |@word trial:9 cu:1 version:2 open:1 propagate:1 jacob:3 initial:2 series:1 exclusively:1 denoting:2 esj:1 current:7 contextual:1 nt:1 nowlan:4 activation:2 motor:1 remove:1 update:1 discrimination:2 selected:1 discovering:1 smith:2 short:1 recherche:1 quantized:1 node:1 sigmoidal:1 five:1 differential:1 become:1 consists:1 fitting:1 introduce:4 operationnelle:1 tlog:1 indeed:3 expected:1 alspector:1 behavior:2 actual:2 considering:2 cardinality:1 becomes:1 begin:1 provided:1 underlying:1 notation:1 what:1 interpreted:5 string:6 watrous:3 jik:1 nj:2 temporal:7 collecting:1 exactly:1 control:1 unit:5 producing:1 positive:1 local:4 consequence:1 mach:1 path:2 might:1 mateo:1 hmms:11 limited:1 mentioning:1 range:1 directed:2 practical:1 implement:1 firenze:1 backpropagation:3 differs:1 procedure:4 bell:2 thought:1 significantly:2 regular:3 suggest:1 symbolic:1 cannot:3 convenience:1 context:1 restriction:1 map:2 missing:1 baum:1 backpropa:1 yt:20 straightforward:1 maximizing:1 automaton:2 focused:1 formulate:1 qc:1 welch:1 simplicity:1 ergodic:1 survey:1 rule:2 retrieve:1 updated:2 target:7 suppose:3 play:1 construction:1 homogeneous:1 vork:1 pa:1 recognition:3 jk:2 predicts:1 labeled:1 bottom:2 observed:1 subproblem:2 role:1 solved:1 worst:2 connected:1 sun:1 ran:1 mozer:2 dempster:2 ui:1 dynamic:3 trained:6 depend:1 learner:1 easily:1 sondhi:1 alphabet:1 train:2 informatique:1 effective:1 describe:2 artificial:1 approached:1 whose:1 modular:3 quite:1 plausible:1 otherwise:1 grammar:11 statistic:1 h3c:1 laird:1 final:4 fsa:5 sequence:20 propose:1 relevant:1 cacciatore:3 representational:1 description:1 convergence:6 extending:1 produce:1 perfect:2 help:1 depending:2 recurrent:15 montreal:1 stat:1 derive:1 measured:2 ij:1 qt:1 zit:1 soc:1 auxiliary:1 kuhn:3 inhomogeneous:1 stochastic:2 successor:2 generalization:3 decompose:1 dipartimento:1 strictly:1 pl:1 extension:1 credit:2 ic:1 considered:2 seed:1 mapping:2 predict:1 viterbi:1 major:1 estimation:4 proc:1 integrates:1 label:1 successfully:1 minimization:1 j7:1 gaussian:1 always:2 rather:1 og:1 derived:1 bernoulli:1 likelihood:9 tech:1 inference:4 xip:1 iohmm:5 hidden:9 relation:1 transformed:1 reproduce:1 interested:1 arg:1 classification:5 flexible:1 among:2 denoted:2 constrained:1 softmax:3 initialize:1 yip:4 mutual:2 special:1 frasconi:13 having:3 extraction:1 represents:1 unsupervised:2 yoshua:5 connectionist:2 np:1 report:4 serious:1 kompe:1 randomly:1 composed:1 intell:1 wijk:1 adjust:1 weakness:1 mixture:5 chain:2 succeeded:1 partial:1 decoupled:1 incomplete:1 old:1 desired:2 pollack:2 minimal:2 column:3 modeling:3 giles:2 markovian:1 tp:2 assignment:2 maximization:1 introducing:1 vertex:1 recognizing:2 successful:3 characterize:1 reported:2 dependency:8 density:1 fundamental:2 nips2:1 probabilistic:5 yl:1 lee:1 moody:1 connectivity:2 squared:1 external:1 expert:3 resort:1 style:3 lull:1 derivative:2 simard:1 account:4 potential:2 de:2 lookup:1 summarized:1 depends:1 root:1 lab:1 portion:1 capability:3 florence:1 minimize:1 oi:1 accuracy:2 kaufmann:4 maximized:1 correspond:1 miller:1 rabiner:1 climbing:1 dealt:2 bayesian:5 mmi:2 mc:2 trajectory:3 worth:1 researcher:1 classified:1 converged:1 touretzky:1 ed:2 uip:3 pp:1 involved:1 universite:1 associated:4 di:2 mi:1 static:5 bridle:2 con:1 proved:1 adjusting:1 iji:1 knowledge:3 ut:19 carefully:1 ea:1 actually:1 appears:1 attained:1 supervised:7 specify:1 response:1 evaluated:1 furthermore:2 rejected:1 until:1 multiscale:1 propagation:2 yf:3 inductive:1 analytically:1 iteratively:1 wp:1 moore:1 deal:2 uniquely:1 recurrence:1 criterion:2 generalized:2 hill:1 evident:1 complete:3 intermittently:1 recently:1 umontreal:1 common:1 multinomial:2 jl:2 discussed:1 interpretation:2 interpret:1 expressing:1 refer:1 automatic:1 similarly:3 pointed:1 language:3 had:1 mealy:1 pnj:1 similarity:2 supervision:1 recognizers:1 italy:1 tesauro:1 store:1 binary:1 rep:1 yi:1 ojlt:1 exploited:1 seen:1 morgan:4 impose:1 determine:1 paradigm:2 ud:1 maximize:1 ii:1 levinson:3 afterwards:1 infer:1 ing:1 smooth:1 technical:1 cross:2 long:6 mle:2 prediction:4 involving:3 converging:1 controller:2 essentially:1 expectation:2 multilayer:1 iteration:1 whereas:2 addition:1 separately:1 interval:1 addressed:1 flammia:1 unlike:1 ascent:2 cowan:2 effectiveness:1 jordan:1 extracting:1 noting:1 presence:1 feedforward:3 bengio:17 intermediate:1 architecture:27 restrict:1 topology:1 reduce:1 iogp:1 prototype:1 qj:1 whether:1 ul:1 effort:1 render:1 speech:2 jj:3 adequate:1 tij:1 zlp:1 iterating:1 useful:1 tune:1 amount:2 processed:2 informatica:1 simplest:1 angluin:2 designer:1 estimated:1 sistemi:1 correctly:2 yy:1 discrete:8 shall:1 paolo:5 independency:2 indefinite:1 backward:3 graph:6 asymptotically:1 inclusively:1 sum:1 fraction:2 compete:1 almost:1 decide:1 holmdel:1 layer:3 ct:1 guaranteed:1 adapted:1 constraint:1 constrain:1 aspect:1 span:1 min:1 jjt:1 structured:2 uf:2 according:2 combination:1 poor:1 belonging:2 em:14 pti:1 wi:2 modification:1 intuitively:1 mori:1 equation:5 mutually:1 etworks:1 previously:1 wellekens:2 end:1 subnetworks:11 available:2 apply:2 generic:2 alternative:3 slower:1 hat:1 top:2 tomita:5 include:1 qit:3 exploit:2 establish:1 universita:1 noticed:2 already:1 question:1 parametric:1 exclusive:1 rt:1 subnetwork:6 gradient:5 link:3 separate:1 hmm:10 me:2 seven:1 fy:3 discriminant:2 trivial:1 induction:3 length:3 modeled:2 difficult:2 subproblems:1 hij:5 favorably:1 adjustable:1 perform:2 gated:1 allowing:1 observation:2 markov:9 arc:1 finite:6 iohmms:4 descent:3 behave:2 extended:1 hinton:1 frame:1 intermittent:1 synchronously:1 introduced:1 pair:1 required:1 specified:4 connection:3 acoustic:1 learned:1 nip:1 trans:2 able:1 dynamical:4 pattern:1 royal:1 memory:2 oj:6 max:1 analogue:1 power:1 suitable:1 ation:1 difficulty:2 natural:1 hybrid:2 predicting:1 indicator:1 bourlard:2 rely:1 improve:1 temporally:1 extract:1 epoch:1 prior:2 ipj:1 fully:2 dsi:1 plant:1 interesting:3 acyclic:1 validation:2 contingency:1 integrate:1 pij:11 propagates:1 principle:2 editor:2 rubin:1 storing:1 share:1 production:3 course:1 summary:1 last:3 aij:1 formal:1 allow:2 taking:1 grammatical:1 feedback:1 transition:11 forward:2 adaptive:2 jump:1 simplified:1 san:1 far:1 transaction:1 sj:3 global:2 sequentially:1 reestimation:1 summing:1 corpus:1 gem:4 assumed:1 xi:1 subsequence:1 continuous:2 un:2 table:5 nature:1 learn:2 ca:2 requested:1 investigated:1 complex:2 did:1 main:1 fashion:1 lc:2 position:1 lie:1 unifi:1 ib:1 admissible:1 removing:1 formula:1 cog:1 xt:26 jt:3 gating:2 symbol:1 restricting:1 sequential:2 effectively:3 conditioned:1 chen:2 depicted:2 lt:2 simply:2 likely:1 expressed:1 scalar:1 corresponds:1 conditional:5 consequently:1 ann:1 absence:1 change:1 specifically:1 lui:2 wt:1 called:2 accepted:1 experimental:2 arbor:1 iipj:1 perceptrons:1 internal:4 support:1 accelerated:1 dept:1 tested:1 phenomenon:1 |
7,204 | 965 | Learning from queries for maximum
information gain in imperfectly learnable
problems
Peter Sollich David Saad
Department of Physics, University of Edinburgh
Edinburgh EH9 3JZ, U.K.
P.Sollich~ed.ac.uk. D.Saad~ed.ac.uk
Abstract
In supervised learning, learning from queries rather than from
random examples can improve generalization performance significantly. We study the performance of query learning for problems
where the student cannot learn the teacher perfectly, which occur
frequently in practice. As a prototypical scenario of this kind, we
consider a linear perceptron student learning a binary perceptron
teacher. Two kinds of queries for maximum information gain, i.e.,
minimum entropy, are investigated: Minimum student space entropy (MSSE) queries, which are appropriate if the teacher space
is unknown, and minimum teacher space entropy (MTSE) queries,
which can be used if the teacher space is assumed to be known, but
a student of a simpler form has deliberately been chosen. We find
that for MSSE queries, the structure of the student space determines the efficacy of query learning, whereas MTSE queries lead
to a higher generalization error than random examples, due to a
lack of feedback about the progress of the student in the way queries
are selected.
1
INTRODUCTION
In systems that learn from examples, the traditional approach has been to study
generalization from random examples, where each example is an input-output pair
288
Peter Sollich, David Saad
with the input chosen randomly from some fixed distribution and the corresponding
output provided by a teacher that one is trying to approximate. However, random
examples contain less and less new information as learning proceeds. Therefore,
generalization performance can be improved by learning from queries, i. e., by choosing the input of each new training example such that it will be, together with its
expected output, in some sense 'maximally useful'. The most widely used measure of 'usefulness' is the information gain, i.e ., the decrease in entropy of the
post-training probability distributions in the parameter space of the student or the
teacher. We shall call the resulting queries 'minimum (student or teacher space)
entropy (MSSE/MTSE) queries'; their effect on generalization performance has recently been investigated for perfectly learnable problems, where student and teacher
space are identical (Seung et al., 1992, Freund et al., 1993, Sollich, 1994), and was
found to depend qualitatively on the structure of the teacher. For a linear perceptron, for example, one obtains a relative reduction in generalization error compared
to learning from random examples which becomes insignificant as the number of
training examples, p, tends to infinity. For a perceptron with binary output, on the
other hand, minimum entropy queries result in a generalization error which decays
exponentially as p increases, a marked improvement over the much slower algebraic
decay with p in the case of random examples.
In practical situations, one almost always encounters imperfectly learnable problems,
where the student can only approximate the teacher, but not learn it perfectly.
Imperfectly learnable problems can arise for two reasons: Firstly, the teacher space
(i.e., the space of models generating the data) might be unknown. Because the
teacher space entropy is then also unknown, MSSE (and not MTSE) queries have
to be used for query learning. Secondly, the teacher space may be known, but a
student of a simpler structure might have deliberately been chosen to facilitate or
speed up training, for example. In this case, MTSE queries could be employed as
an alternative to MSSE queries. The motivation for doing this would be strongest
if, as in the learning scenario that we consider below, it is known from analyses
of perfectly learnable problems that the structure of the teacher space allows more
significant improvements in generalization performance from query learning than
the structure of the student space.
With the above motivation in mind, we investigate in this paper the performance
of both MSSE and MTSE queries for a prototypical imperfectly learnable problem, in which a linear perceptron student is trained on data generated by a binary
perceptron teacher. Both student and teacher are specified by an N-dimensional
weight vector with real components, and we will consider the thermodynamic limit
N -+ 00, p -+ 00,0:' = piN = const. In Section 2 below we calculate the generalization error for learning from random examples. In Sections 3 and 4 we compare
the result to MSSE and MTSE queries. Throughout, we only outline the necessary calculations; for details, we refer the reader to a forthcoming publication. We
conclude in Section 5 with a summary and brief discussion of our results.
2
LEARNING FROM RANDOM EXAMPLES
We denote students and teachers by N (for 'Neural network') and V (for 'element
of the Version space', see Section 4), respectively, and their corresponding weight
Learning from Queries for Maximum Information Gain
289
vectors by W Nand w v . For an input vector x, the outputs of a given student and
teacher are
YN =
XTWN' Yv = sgn(
xTwv) .
7N
7N
Assuming that inputs are drawn from a uniform distribution over the hypersphere
x 2 = N, and taking as our error measure the standard squared output difference
(YN - Yv)2, the generalization error, i. e., the average error between student Nand
teacher V when tested on random test inputs, is given by
!
?g(N,
V)
1[
= '2
QN
(2) 1/2] ,
R
+ 1- 2..ftTv;:
(1)
where we have set R = ];WrrWv,QN = ];w~,Qv = ];w~.
As our training algorithm we take stochastic gradient descent on the trammg
error E t , which for a training set e(p) = {( x~ ,Y~
Yv( x~)), J.l
1 ... p} is
Et
! 2:~(Y~ - YN(X~))2. A weight decay term !AW~ is added for regularization, i.e., to prevent overfitting. Stochastic gradient descent on the resulting energy function E = E t + !AW~ yields a Gibbs post-training distribution of students,
p(wNle(p)) ex exp(-E/T), where the training temperature T measures the amount
of stochasticity in the training algorithm. For the linear perceptron students considered here, this distribution is Gaussian, with covariance matrix TM,N-l, where
(IN denotes the N x N identity matrix)
=
=
MN = AlN
=
+ ~ 2:~=1 x~(x~f?
Since the length ofthe teacher weight vector Wv does not affect the teacher outputs,
we assume a spherical prior on teacher space, P(w v ) ex 6(w~-N), for which Qv = 1.
Restricting attention to the limit of zero training temperature, it is straightforward
to calculate from eq . (1) the average generalization error obtained by training on
random examples
?g - ?g,min
=;:1
[ AoptG
+ A(Aopt - A) OG]
OA '
(2)
with the function G = (]; tr M,N-1) P( {xl'}) given by (Krogh and Hertz, 1992)
G=
2\ [1 - a - A+ )(1- a - A)2 + 4A] .
(3)
In eq. (2) we have explicitly subtracted the minimum achievable generalization error,
= !(1-2/11'), which is nonzero since a linear perceptron cannot approximate a
binary percept ron perfectly. At finite a, the generalization error is minimized when
the weight decay is set to its optimal value A = Aopt
11'/2 - 1. Note that since
both G and OG/OA tend to zero as a --+ 00, the generalization error for random
examples approaches the minimum achievable generalization error in this limit.
?g,min
=
3
MINIMUM STUDENT SPACE ENTROPY QUERIES
We now calculate the generalization performance resulting from MSSE queries. For
the training algorithm introduced in the last section, the student space entropy
(normalized by N) is given by
290
Peter Sollich, David Saad
""T""----r----r---------------,
3.0
- - A== 0.01
---- A == 0.1
-------- A == 1
2.5
2.0
1.5
-------------------::~-::-=.:--~~.J
.. _-----------1.0-+--.::..::......-.-----,----.-----r-----I
3
o
4
5
2
a
Figure 1: Relative improvement
weight decay A = 0.01,0 .1,1.
K,
in generalization error due to MSSE queries, for
1
S'" = - 2N lndet M N
,
where we have omitted an unimportant constant which depends on the training
temperature only. This entropy is minimized by choosing each new query along
the direction corresponding to the minimal eigenvalue of the existing MN (Sollich,
1994). The expression for the resulting average generalization error is given by
eq. (2) with G replaced by its analogue for MSSE queries (Sollich, 1994)
~a
1- ~a
GQ +----:- A + [a] + 1 A + [a] ,
where [a] is the greatest integer less than or equal to a and ~o' = a-raj. We define
the improvement factor K, as the ratio of the generalization error (with the minimum
achievable generalization error subtracted as in eq. (2)) for random examples to that
for MSSE queries. Figure 1 shows K(O') for several values of the weight decay A.
Comparing with existing results (Sollich, 1994), we find that K, is exactly the same
as if our linear student were trying to approximate a linear teacher with additive
noise of variance Aopt on the outputs. For large a, one can show (Sollich, 1994) that
K, = 1 + I/O' + O(I/a 2 ) and hence the relative reduction in generalization error due
to querying tends to zero as a - I - 00. We investigate in the next section whether it
is possible to improve generalization performance more significantly by using MTSE
queries.
4
MINIMUM TEACHER SPACE ENTROPY QUERIES
We now consider the generalization performance achieved by MTSE queries. We
remind readers that such queries could be used if the teacher space is known, but
a student of a simpler functional form has deliberately chosen. The aim in using
MTSE rather than MSSE queries would be to exploit the structure of the teacher
space if this is known (for perfectly learnable problems) to make query learning very
efficient compared to random examples.
For the case of noise free training data under consideration, the posterior probability
distribution in teacher space given a certain training set is proportional to the prior
291
Learning from Queries for Maximum Information Gain
distribution on the version space (the set of all teachers that could have produced
the training set without error) and zero everywhere else. From this the (normalized)
teacher space entropy can be derived to be, up to an additive constant,
1
Sv = N In V(p),
where the version space volume V(p) is given by (8(z) = 1 for z
V(p) =
> 0 and 0 otherwise)
JdwvP(w v) n~=l 8(JNy/Jw~x/J).
It can easily be verified that this entropy is minimized l by choosing queries x which
(bisect' the existing version space, i. e., for which the hyperplane perpendicular to
x splits the version space into two equal halves (Seung et al., 1992, Freund et al.,
1993). Such queries lead to an exponentially shrinking version space, V(p) = 2- P ,
and hence a linear decrease of the entropy, Sv = -a In 2. We consider instead
queries which achieve qualitatively the same effect, but permit a much simpler
analysis of the resulting student performance. They are similar to those studied in
the context of a learnable problem by Watkin and Rau (1992), and are defined as
follows . The (p + 1)th query is obtained by first picking a random teacher vector
wp from the version space defined by the existing p training examples, and then
picking the new training input Xp+l from the distribution of random inputs but
under the constraint that x;+1wp = O.
For the calculation of the student performance, i. e., the average generalization error,
achieved by the approximate MTSE queries described above, we use an approximation based on the following observation. As the number of training examples,
p, increases, the teacher vectors wp from the version space will align themselves
with the true teacher w~; their components along the direction of w~ will increase,
whereas their components perpendicular to we will decrease, varying widely across
the N - 1 dimensional hyperplane perpendicular to we . Following Watkin and Rau
(1992), we therefore assume that the only significant effect of choosing queries xp+1
with X;+lWP = 0 is on the distribution of the component ofxp+l along we . Writing
this component as x~+1 = x;+1 w~/lwel, its probability distribution can readily be
shown to be
P(x~+1) ex: exp (-~(x~+1lsp)2) ,
(4)
where sp is the sine of the angle between w p and we. For finite N, the value of sp
is dependent on the p previous training examples that define the existing version
space and on the teacher vector wp sampled randomly from this version space. In
the thermodynamic limit, however, the variations of sp become vanishingly small
and we can thus replace sp by its average value, which is a function of palone.
In the thermodynamic limit, this average value becomes a continuous function of
a = pIN , the number of training examples per weight, which we denote simply by
sea) . The calculation can then be split into two parts: First, the function sea) is
obtained from a calculation of the teacher space entropy using the replica method,
generalizing the results of Gyorgi and Tishby (1990) . The average generalization
1 More precisely, what is minimized is the value of the entropy after a new training
example (x, y) is added, averaged over the distribution of the unknown new training output
y given the new training input x and the existing training set; see Sollich (1994).
292
Peter Sollich, David Saad
0
................?, .......,.,.....,.
-1
--':''''''::-.::, -............
-----------
-2
.....
'
-3
.
...?.....??......
-4
-5
0
2
3
4
5 a 6
7
8
9
10
Figure 2: MTSE queries: Teacher space entropy, Sv (with value for random examples plotted for comparison), and In s, the log of the sine of the angle between the
true teacher and a random teacher from the version space.
error can then be calculated by using an extension of the response function method
described in (Sollich, 1994b) or by another replica calculation (now in student space)
as in (Dunmur and Wallace, 1993).
Figure 2 shows the effects of (approximate) MTSE queries in teacher space. For large
a values, the teacher space entropy decreases linearly with a, with gradient c : : : : 0.44,
whereas the entropy for random examples , also shown for comparison , decreases
much more slowly (asymptotically like -In a, see (Gyorgi and Tishby, 1990)). The
linear a-dependence of the entropy for queries corresponds to an average reduction
of the version space volume with each new training example by a factor of exp( -c) : : : :
0.64, which is reasonably close to the factor ~ for proper bisection of the version
space. This justifies our choice of analysing approximate MTSE queries rather than
true MTSE queries, since the former achieve qualitatively the same results as the
latter.
Before discussing the student performance achieved by (approximate) MTSE
queries, we note from figure 2 that In s( a) decreases linearly with a for large a,
with the same gradient as the teacher space entropy. Hence s( a) ex: exp( -ca) for
large a, and MTSE queries force the average teacher from the version space to
approach the true teacher exponentially quickly. It can easily be shown that if we
were learning with a binary perceptIOn student, i. e., if the problem were perfectly
learnable, then this would result in an exponentially decaying generalization error,
?g ex: exp( -ca). MTSE queries would thus lead to a marked improvement in generalization performance over random examples (for which ?g ex: l/a, see (Gyorgi and
Tishby, 1990)). It is this significant benefit (in teacher space) of query learning that
provides the motivation for using MTSE queries in imperfectly learnable problems
such as the one considered here.
The results plotted in Figure 3 for the average generalization error achieved by the
linear perceptron student show, however, that MTSE queries do not have the desired effect. Far from translating the benefits in teacher space into improvements
in generalization performance for the linear student, they actually lead to a deterioration of generalization performance, i. e., a larger generalization error than that
Learning from Queries for Maximum Information Gain
- - >. = 0.01
---- >. = 0.1
-------- >. = 1
0.6
0.5
fg
...\ - ,
"
0.4
,
'..". ,,
'......., ..
0.3
--------
- ---- --------......-=---=--=---=--d--
-~
0.2
o
293
2
3
4
5 a 6
7
8
9
10
Figure 3: Generalization error for MTSE queries (higher curves of each pair) and
random examples (lower curves), for weight decay>. = 0.01,0.1,1. The curves for
random examples (which are virtually indistinguishable from one another already
at a = 10) converge to the minimum achievable generalization error fg,min (dotted
line) as a -+ 00.
obtained for random examples. Worse still, they 'mislead' the student to such an
extent that the minimum achievable generalization error is not reached even for an
infinite number of training examples, a -+ 00. How does this happen? It can be
verified that the angle between the student and teacher weight vectors tends to zero
for a -+ 00 as expected, while QN, the normalized squared length of the student
weight vector, approaches
QN(a
-+ 00)
S
= -2 ( ---=
11'
>. + s2
)2 ,
(5)
where s = Jooo da s( a), s2 = Jooo da s2 (a). Unless the weight decay parameter >.
happens to be equal to s - s2, this is different from the optimal asymptotic value,
which is 2/11'. This is the reason why in general the linear student does not reach
the minimum possible generalization error even as a -+ 00. The approach of QN
to its non-optimal asymptotic value can cause an increase in the generalization
error for large a and a corresponding minimum of the generalization error at some
finite a, as can be seen in the plots for>. = 0.01 and 0.1 in Figure 3. For>. = 0,
eq. (5) has the following intuitive interpretation: As a increases, the version space
shrinks around the true teacher w~, and hence MTSE queries become 'more and
more orthogonal' to w~ . As a consequence, the distribution of training inputs along
the direction of
is narrowed down progressively (compare eq. (4)). Trying to
find a best fit to the teacher's binary output function over this narrower range of
inputs, the linear student learns a function which is steeper than the best fit over
the range of random inputs (which would give minimum generalization error). This
corresponds to a suboptimally large length of the student weight vector in agreement
with eq. (5): QN(a -+ 00) > 2/11' for>. = a because s2 < s.
we
Summarizing the results of this section, we have found that although MTSE queries
are very beneficial in teacher space, they are entirely misleading for the linear student, to the extent that the student does not learn to approximate the teacher
optimally even for an infinite number of training examples.
294
5
Peter Sollich, David Saad
SUMMARY AND DISCUSSION
We have found in our study of an imperfectly learnable problem with a linear student
and a binary teacher that queries for minimum student and teacher space entropy,
respectively, have very different effects on generalization performance. Minimum
student space entropy (MSSE) queries essentially have the same effect as for a linear
student learning a noisy linear teacher, apart from a nonzero minimum value of the
generalization error due to the unlearnability of the problem . Hence the structure
of the student space is the dominating influence on the efficacy of query learning.
Minimum teacher space entropy queries (MTSE), on the other hand, perform worse
than random examples, leading to a higher generalization error even for an infinite
number of training examples. With the benefit of hindsight, we note that this makes
intuitive sense since the teacher space entropy, according to which MTSE queries
are selected, contains no feedback about the progress of the student in learning the
required generalization task, and thus MTSE queries cannot be guaranteed to have
a positive effect.
Our results, then, are a mixture of good and bad news for query learning for maximum information gain in imperfectly learnable problems: The bad news is that
MTSE queries, due to a lack of feedback information about student progress, are
not enough to translate significant benefits in teacher space into similar improvements of student performance and may in fact yield worse performance than random
examples. The good news is that for MSSE queries, we have found evidence that
the structure of the student space is the key factor in determining the efficacy of
query learning. If this result holds more generally, then statements about the benefits of query learning can be made on the basis of how one is trying to learn only,
independently of what one is trying to learn-a result of great practical significance.
References
A P Dunmur and D J Wallace (1993). Learning and generalization in a linear
perceptron stochastically trained with noisy data. J. Phys. A, 26:5767-5779.
Y Freund, H S Seung, E Shamjr, and N Tishby (1993). Information, prediction,
and query by committee. In S J Hanson, J D Cowan, and C Lee Giles, editors,
NIPS 5, pages 483-490, San Mateo, CA, Morgan Kaufmann.
G Gyorgi and N Tishby (1990). Statistical theory of learning a rule. In
W Theumann and R Koberle, editors, Neu.ral Networks and Spin Glasses, pages
3-36. Singapore , World Scientific.
A Krogh and J A Hertz (1992). Generalization in a linear perceptron in the presence
of noise. J. Phys. A,25:1135-1147.
P Sollich (1994). Query construction , entropy, and generalization in neural network
models. Phys. Rev. E,49:4637-4651.
P Sollich (1994b) . Finite-size effects in learning and generalization in linear perceptrons. J. Phys. A, 27:7771-7784.
H S Seung, M Opper, and H Sompolinsky (1992) . Query by committee. In Proceedings of COLT '92, pages 287-294, New York, ACM .
T L H Watkin and A Rau (1992). Selecting examples for perceptrons. J . Phys. A,
25:113-121.
| 965 |@word version:15 achievable:5 covariance:1 tr:1 reduction:3 contains:1 efficacy:3 selecting:1 existing:6 comparing:1 readily:1 additive:2 happen:1 plot:1 progressively:1 half:1 selected:2 hypersphere:1 provides:1 ron:1 firstly:1 simpler:4 along:4 become:2 expected:2 themselves:1 frequently:1 wallace:2 spherical:1 gyorgi:4 becomes:2 provided:1 what:2 kind:2 aln:1 hindsight:1 exactly:1 uk:2 theumann:1 yn:3 before:1 positive:1 tends:3 limit:5 consequence:1 might:2 studied:1 mateo:1 perpendicular:3 range:2 averaged:1 practical:2 practice:1 lndet:1 significantly:2 cannot:3 close:1 context:1 influence:1 writing:1 unlearnability:1 straightforward:1 attention:1 independently:1 mislead:1 rule:1 variation:1 construction:1 agreement:1 element:1 calculate:3 news:3 sompolinsky:1 decrease:6 bisect:1 lwp:1 seung:4 trained:2 depend:1 basis:1 easily:2 query:70 choosing:4 widely:2 larger:1 dominating:1 otherwise:1 noisy:2 eigenvalue:1 gq:1 vanishingly:1 translate:1 achieve:2 intuitive:2 sea:2 generating:1 ac:2 progress:3 eq:7 krogh:2 direction:3 stochastic:2 sgn:1 translating:1 msse:14 generalization:46 secondly:1 extension:1 hold:1 around:1 considered:2 exp:5 great:1 omitted:1 trammg:1 qv:2 always:1 gaussian:1 aim:1 rather:3 og:2 varying:1 publication:1 derived:1 improvement:7 ral:1 dunmur:2 sense:2 summarizing:1 glass:1 dependent:1 nand:2 colt:1 equal:3 identical:1 minimized:4 randomly:2 replaced:1 investigate:2 mixture:1 necessary:1 orthogonal:1 unless:1 desired:1 plotted:2 minimal:1 giles:1 imperfectly:7 uniform:1 usefulness:1 tishby:5 optimally:1 teacher:56 aw:2 sv:3 jooo:2 lee:1 physic:1 picking:2 together:1 quickly:1 squared:2 slowly:1 watkin:3 worse:3 stochastically:1 leading:1 student:47 explicitly:1 depends:1 sine:2 doing:1 steeper:1 reached:1 yv:3 decaying:1 narrowed:1 spin:1 variance:1 kaufmann:1 percept:1 yield:2 ofthe:1 produced:1 bisection:1 strongest:1 reach:1 phys:5 ed:2 neu:1 energy:1 gain:7 sampled:1 actually:1 higher:3 supervised:1 response:1 improved:1 maximally:1 jw:1 shrink:1 hand:2 lack:2 scientific:1 facilitate:1 effect:9 contain:1 normalized:3 true:5 deliberately:3 former:1 regularization:1 hence:5 nonzero:2 wp:4 indistinguishable:1 trying:5 outline:1 temperature:3 consideration:1 recently:1 lsp:1 functional:1 exponentially:4 volume:2 interpretation:1 significant:4 refer:1 rau:3 gibbs:1 stochasticity:1 align:1 posterior:1 raj:1 apart:1 scenario:2 certain:1 wv:1 binary:7 discussing:1 seen:1 minimum:19 morgan:1 employed:1 converge:1 thermodynamic:3 calculation:5 post:2 prediction:1 essentially:1 deterioration:1 achieved:4 whereas:3 else:1 saad:6 tend:1 virtually:1 cowan:1 call:1 integer:1 presence:1 split:2 enough:1 affect:1 fit:2 forthcoming:1 perfectly:7 tm:1 whether:1 expression:1 peter:5 algebraic:1 york:1 cause:1 useful:1 generally:1 unimportant:1 amount:1 singapore:1 dotted:1 per:1 shall:1 key:1 drawn:1 prevent:1 verified:2 replica:2 asymptotically:1 angle:3 everywhere:1 aopt:3 almost:1 throughout:1 reader:2 eh9:1 entirely:1 guaranteed:1 occur:1 infinity:1 constraint:1 precisely:1 speed:1 min:3 department:1 according:1 hertz:2 across:1 sollich:15 beneficial:1 rev:1 happens:1 pin:2 committee:2 mind:1 permit:1 appropriate:1 subtracted:2 alternative:1 encounter:1 slower:1 denotes:1 const:1 exploit:1 added:2 already:1 dependence:1 traditional:1 gradient:4 oa:2 extent:2 reason:2 assuming:1 length:3 suboptimally:1 remind:1 ratio:1 statement:1 proper:1 unknown:4 perform:1 observation:1 finite:4 descent:2 situation:1 david:5 introduced:1 pair:2 required:1 specified:1 hanson:1 nip:1 proceeds:1 below:2 perception:1 analogue:1 greatest:1 force:1 mn:2 improve:2 misleading:1 brief:1 koberle:1 prior:2 determining:1 relative:3 asymptotic:2 freund:3 prototypical:2 proportional:1 querying:1 xp:2 editor:2 summary:2 last:1 free:1 perceptron:11 taking:1 fg:2 edinburgh:2 benefit:5 feedback:3 calculated:1 curve:3 world:1 opper:1 qn:6 qualitatively:3 made:1 san:1 far:1 approximate:9 obtains:1 overfitting:1 assumed:1 conclude:1 continuous:1 why:1 jz:1 learn:6 reasonably:1 ca:3 investigated:2 da:2 sp:4 significance:1 linearly:2 motivation:3 noise:3 arise:1 s2:5 shrinking:1 xl:1 learns:1 down:1 bad:2 learnable:12 insignificant:1 decay:8 evidence:1 restricting:1 justifies:1 entropy:26 generalizing:1 simply:1 corresponds:2 determines:1 acm:1 marked:2 identity:1 narrower:1 replace:1 analysing:1 infinite:3 hyperplane:2 perceptrons:2 latter:1 tested:1 ex:6 |
7,205 | 966 | Boltzmann Chains and Hidden
Markov Models
Lawrence K. Saul and Michael I. Jordan
lksaulOpsyche.mit.edu, jordanOpsyche.mit.edu
Center for Biological and Computational Learning
Massachusetts Institute of Technology
79 Amherst Street, E10-243
Cambridge, MA 02139
Abstract
We propose a statistical mechanical framework for the modeling
of discrete time series. Maximum likelihood estimation is done via
Boltzmann learning in one-dimensional networks with tied weights.
We call these networks Boltzmann chains and show that they
contain hidden Markov models (HMMs) as a special case. Our
framework also motivates new architectures that address particular shortcomings of HMMs. We look at two such architectures:
parallel chains that model feature sets with disparate time scales,
and looped networks that model long-term dependencies between
hidden states. For these networks, we show how to implement
the Boltzmann learning rule exactly, in polynomial time, without
resort to simulated or mean-field annealing. The necessary computations are done by exact decimation procedures from statistical
mechanics.
1
INTRODUCTION AND SUMMARY
Statistical models of discrete time series have a wide range of applications, most
notably to problems in speech recognition (Juang & Rabiner, 1991) and molecular
biology (Baldi, Chauvin, Hunkapiller, & McClure, 1992). A common problem in
these fields is to find a probabilistic model, and a set of model parameters, that
436
Lawrence K. Saul, Michael I. Jordan
account for sequences of observed data. Hidden Markov models (HMMs) have been
particularly successful at modeling discrete time series. One reason for this is the
powerful learning rule (Baum) 1972?) a special case of the Expectation- Maximization
(EM) procedure for maximum likelihood estimation (Dempster) Laird) & Rubin)
1977).
In this work) we develop a statistical mechanical framework for the modeling of
discrete time series. The framework enables us to relate HMMs to a large family
of exactly solvable models in statistical mechanics. The connection to statistical
mechanics was first noticed by Sourlas (1989?) who studied spin glass models of
error-correcting codes. We view the estimation procedure for HMMs as a special
(and particularly tractable) case of the Boltzmann learning rule (Ackley) Hinton) &
Sejnowski) 1985; Byrne) 1992).
The rest of this paper is organized as follows . In Section 2) we review the modeling
problem for discrete time series and establish the connection between HMMs and
Boltzmann machines. In Section 3) we show how to quickly determine whether
or not a particular Boltzmann machine is tractable) and if so) how to efficiently
compute the correlations in the Boltzmann learning rule. Finally) in Section 4)
we look at two architectures that address particular weaknesses of HMMs: the
modelling of disparate time scales and long-term dependencies.
2
MODELING DISCRETE TIME SERIES
A discrete time series is a sequence of symbols {jdr=l in which each symbol belongs
to a finite countable set) i.e. jl E {1) 2) .. . ) m}. Given one long sequence) or perhaps
many shorter ones) the modeling task is to characterize the probability distribution
from which the time series are generated.
2.1
HIDDEN MARKOV MODELS
A first-order Hidden Markov Model (HMM) is characterized by a set of n hidden
states) an alphabet of m symbols) a transmission matrix ajj') an emission matrix
bjj ) and a prior distribution 7I'j over the initial hidden state. The sequence of states
{idr=l and symbols {jdr=l is modeled to occur with probability
(1)
The modeling problem is to find the parameter values (ajj' , bij ) 7I'j) that maximize
the likelihood of observed sequences of training data. We will elaborate on the
learning rule in section 2.3) but first let us make the connection to a well-known
family of stochastic neural networks , namely Boltzmann machines.
2.2
BOLTZMANN MACHINES
Consider a Boltzmann machine with m-state visible units) n-state hidden units) tied
weights) and the linear architecture shown in Figure 1. This example represents the
simplest possible Boltzmann "chain))) one that is essentially equivalent to a firstorder HMM unfolded in time (MacKay) 1994). The transition weights Aii' connect
adjacent hidden units) while the emission weights Bj j connect each hidden unit to
Boltzmann Chains and Hidden Markov Models
437
1J
Bij
ViS~ble
un~ts
Bij
hidden
units
? ? ? A i i , Au'
Figure 1: Boltzmann chain with n-state hidden units, m-state visible units, transition weights Aiil, emission weights Bij, and boundary weights IIi.
its visible counterpart. In addition, boundary weights IIi model an extra bias on
the first hidden unit. Each configuration of units represents a state of energy
L-1
1t[{il' jd] =
-Ilil -
L
L
Ailil+t -
l=l
2:
Bitio
(2)
l=l
where {idf=l ({jl }f=l) is the sequence of states over the hidden (visible) units. The
probability to find the network in a particular configuration is given by
' '}) = Ze
1 -{31-l ,
P({ Zl,)l
where f3
(3)
= I/T is the inverse temperature, and the partition function
L
Z =
e-fJ'H.
(4)
{idd
is the sum over states that normalizes the Boltzmann distribution, eq. (3).
Comparing this to the HMM distribution, eq. (1), it is clear that any first-order
HMM can be represented by the Boltzmann chain of figure 1, provided we take 1
IIi = TIn 7ri?
(5)
Later, in Section 4, we will consider more complicated chains whose architectures
address particular shortcomings of HMMs. For now, however, let us continue to
develop the example of figure 1, making explicit the connection to HMMs.
Aii'
2.3
= TIn aij/,
Bij
= TIn bij ,
LEARNING RULES
In the framework of Boltzmann learning (Williams & Hinton, 1990), the data for
our problem consist of sequences of states over the visible units; the goal is to find
the weights (Ail, B ij , IIi) that maximize the likelihood of the observed data. The
likelihood of a sequence {jd is given by the ratio
.
P({il,jd)
P({Jd) = P({idl{jl})
e-{3'H./Z
Zc
= e-{3'H./Zc = Z'
(6)
1 Note, however, that the reverse statement-that for any set of parameters, this Boltzmann chain can be represented as an HMM-is not true. The weights in the Boltzmann
chain represent arbitrary energies between ?oo, whereas the HMM parameters represent
probabilities that are constrained to obey sum rules, such as Lil aiil = 1. The Boltzmann
chain of figure 1 therefore has slightly more degrees of freedom than a first-order HMM.
An interpretation of these extra degrees of freedom is given by MacKay (1994).
438
Lawrence K. Saul, Michael I. Jordan
where Zc is the clamped partition function
Zc =
L e-/3
1i .
(7)
{it}
Note that the sum in Zc is only over the hidden states in the network, while the
visible states are clamped to the observed values bt}.
The Boltzmann learning rule adjusts the weights of the network by gradient-ascent
on the log-likelihood. For the example of figure 1, this leads to weight updates
L-l
~Aii'
=
7J/3
L [(6iil6ilil+Jc -
(8)
(6iil6ilil+l)] ;
l=1
L
~Bij
7J/3
L [(6iil6jjl)C -
(9)
(6i il 6jjl)] ,
l=1
~ni
7J/3 [(6ii1 )c - (6ii1 )] ,
(10)
where 6ij stands for the Kronecker delta function, 7J is a learning rate, and (-) and
(-) c denote expectations over the free and clamped Boltzmann distributions.
The Boltzmann learning rule may also be derived as an Expectation-Maximization
(EM) algorithm. The EM procedure is an alternating two-step method for maximum likelihood estimation in probability models with hidden and observed variables. For Boltzmann machines in general, neither the E-step nor the M-step can
be done exactly; one must estimate the necessary statistics by Monte Carlo simulation (Ackley et al., 1985) or mean-field theory (Peterson & Anderson, 1987).
In certain special cases (e.g. trees and chains) , however, the necessary statistics
can be computed to perform an exact E-step (as shown below). While the Mstep in these Boltzmann machines cannot be done exactly, the weight updates can
be approximated by gradient descent. This leads to learning rules in the form of
eqs. (8-10).
HMMs may be viewed as a special case of Boltzmann chains for which both the
E-step and the M-step are analytically tractable. In this case, the maximization in
the M-step is performed subject to the constraints 2:i e/3 Il ? = 1, 2:il e/3 A ;;1 = 1, and
2:j e/3 B ;i = 1. These constraints imply Z = 1 and lead to closed-form equations
for the weight updates in HMMs.
3
EXACT METHODS FOR BOLTZMANN LEARNING
The key technique to compute partition functions and correlations in Boltzmann
chains is known as decimation. The idea behind decimation 2 is the following. Consider three units connected in series, as shown in Figure 2a. Though not directly
connected, the end units have an effective interaction that is mediated by the middle
one. In fact, the two weights in series exert the same influence as a single effective
weight, given by
(11)
jl
2A
related method, the transfer matrix, is described by Stolarz (1994).
Boltzmann Chains and Hidden Markov Models
439
A~~)'
1.1.
+
A~~).
11.
(b)
(a)
(c)
Figure 2: Decimation, pruning, and joining in Boltzmann machines.
Replacing the weights in this way amounts to integrating out, or decimating, the
degree offreedom represented by the middle unit. An analogous rule may be derived
for the situation shown in Figure 2b. Summing over the degrees of freedom of the
dangling unit generates an effective bias on its parent, given by
ef3 B ?
=
L:: ef3B ?
j ?
(12)
j
We call this the pruning rule. Another type of equivalence is shown in Figure 2c.
The two weights in parallel have the same effect as the sum total weight
Ajjl = A~P + A~i) .
(13)
We call this the joining rule. It holds trivially for biases as well as weights.
The rules for decimating, pruning, and joining have simple analogs in other types
of networks (e.g. the law for combining resistors in electric circuits), and the strategy for exploiting them is a familiar one. Starting with a complicated network,
we iterate the rules until we have a simple network whose properties are easily
computed. A network is tractable for Boltzmann learning if it can be reduced to
any pair of connected units. In this case, we may use the rules to compute all the
correlations required for Boltzmann learning. Clearly, the rules do not make all networks tractable; certain networks (e.g. trees and chains), however, lend themselves
naturally to these types of operations.
4
DESIGNER NETS
The rules in section 3 can be used to quickly assess whether or not a network is
tractable for Boltzmann learning. Conversely, they can be used to design networks
that are computationally tractable. This section looks at two networks designed to
address particular shortcomings of HMMs.
4.1
PARALLEL CHAINS AND DISPARATE TIME SCALES
An important problem in speech recognition (Juang et al., 1991) is how to "combine
feature sets with fundamentally different time scales." Spectral parameters, such
440
Lawrence K. Saul, Michael I. Jordan
fast
features
coupled
hidden
units
slow
features
Figure 3: Coupled parallel chains for features with different time scales.
as the cepstrum and delta-cepstrum, vary on a time scale of 10 msec; on the other
hand, prosodic parameters, such as the signal energy and pitch, vary on a time scale
of 100 msec. A model that takes into account this disparity should avoid two things.
The first is redundancy-in particular, the rather lame solution of oversampling the
nonspectral features. The second is overfitting. How might this arise? Suppose we
have trained two separate HMMs on sequences of spectral and prosodic features,
knowing that the different features "may not warrant a single, unified Markov chain"
(Juang et al., 1991). To exploit the correlation between feature sets, we must now
couple the two HMMs. A naive solution is to form the Cartesian product of their
hidden state spaces and resume training. Unfortunately, this results in an explosion
in the number of parameters that must be fit from the training data. The likely
consequences are overfitting and poor generalization.
Figure 3 shows a network for modeling feature sets with disparate time scales-in
this case, a 2: 1 disparity. Two parallel Boltzmann chains are coupled by weights
that connect their hidden units. Like the transition and emission weights within
each chain, the coupling weights are tied across the length of the network. Note
that coupling the time scales in this way introduces far fewer parameters than
forming the Cartesian product of the hidden state spaces. Moreover, the network is
tractable by the rules of section 3. Suppose, for example, that we wish to compute
the correlation between two neighboring hidden units in the middle of the network.
This is done by first pruning all the visible units, then repeatedly decimating hidden
units from both ends of the network.
Figure 4 shows typical results on a simple benchmark problem, with data generated
by an artificially constructed HMM. We tested the parallel chains model on 10
training sets, with varying levels of built-in correlation between features. A twostep method was used to train the parallel chains. First, we set the coupling weights
to zero and trained each chain by a separate Baum-Welch procedure. Then, after
learning in this phase was complete, we lifted the zero constraints and resumed
training with the full Boltzmann learning rule. The percent gain in this second
phase was directly related to the degree of correlation built into the training data,
suggesting that the coupling weights were indeed capturing the correlation between
feature sets. We also compared the performance of this Boltzmann machine versus
that of a simple Cartesian-product HMM trained by an additional Baum-Welch
procedure. While in both cases the second phase of learning led to reduced training
error, the Cartesian product HMMs were decidedly more prone to overfitting.
Boltzmann Chains and Hidden Markov Models
I
?1500
441
I-J"anv::luu::a:cnu:nnuI I II JOU"'XJ:o:x:o:a:J::)
/tfA='=u"na"m.....,m"l
1-1rainl"O
!
- - croaa-vaJdation
~ 20
'"
'"
10
-1700
200
eoo
400
eoo
0.2
epoch
(a)
0.6
0.'
feature colT8latkwl
0.8
(b)
Figure 4: (a) Log-likelihood versus epoch for parallel chains with 4-state hidden
units, 6-state visible units, and 100 hidden-visible unit pairs (per chain) . The
second jump in log-likelihood occurred at the onset of Boltzmann learning (see
text). (b) Percent gain in log-likelihood versus built-in correlation between feature
sets.
4.2
LOOPS AND LONG-TERM DEPENDENCIES
Another shortcoming of first-order HMMs is that they cannot exhibit long-term
dependencies between the hidden states (Juang et aL , 1991). Higher-order and
duration-based HMMs have been used in this regard with varying degrees of success. The rules of section 3 suggest another approach-namely, designing tractable
networks with limited long-range connectivity. As an example, Figure 5a shows a
Boltzmann chain with an internal loop and a long-range connection between the
first and last hidden units. These extra features could be used to enforce known
periodicities in the time series. Though tractable for Boltzmann learning, the loops
in this network do not fit naturally into the framework of HMMs . Figure 5b shows
learning curves for a toy problem, with data generated by another looped network.
Carefully chosen loops and long-range connections provide additional flexibility in
the design of probabilistic models for time series. Can networks with these extra
features capture the long-term dependencies exhibited by real data? This remains
an important issue for future research .
Acknowledgements
We thank G. Hinton, D. MacKay, P. Stolorz, and C. Williams for useful discussions. This work was funded by ATR Human Information Processing Laboratories ,
Siemens Corporate Research, and NSF grant CDA-9404932 .
References
D. H. Ackley, G . E. Hinton, and T . J. Sejnowski. (1985) A Learning Algorithm for
Boltzmann Machines. Cog. Sci. 9: 147- 160.
P. Baldi, Y. Chauvin , T . Hunkapiller, and M. A. McClure. (1992) Proc. Nat . Acad.
Sci. (USA) 91: 1059-1063.
Lawrence K. Saul, Michael I. Jordan
442
I-tralnin
~
g
crosa..validation
I
12
1.
?700
o
(a)
8
epoch
10
(b)
Figure 5: (a) Looped network. (b) Log-likelihood versus epoch for a looped network
with 4-state hidden units, 6-state visible units, and 100 hidden-visible unit pairs .
L. Baum. (1972) An Inequality and Associated Maximization Technique in Statistical Estimation of Probabilistic Functions of Markov Processes, Inequalities 3:1-8.
Byrne, W. (1992) Alternating Minimization and Boltzmann Machine Learning.
IEEE Trans. Neural Networks 3:612-620.
A. P. Dempster, N. M. Laird, and D. B. Rubin . (1977) Maximum Likelihood from
Incomplete Data via the EM Algorithm . J. Roy. Statist. Soc. B, 39 :1-38 .
C. Itzykson and J . Drouffe. (1991) Statistical Field Theory, Cambridge: Cambridge
University Press.
B. H. Juang and L. R. Rabiner. (1991) Hidden Markov Models for Speech Recognition, Technometrics 33: 251-272.
D. J. MacKay. (1994) Equivalence of Boltzmann Chains and Hidden Markov Models, submitted to Neural Compo
C. Peterson and J. R. Anderson. (1987) A Mean Field Theory Learning Algorithm
for Neural Networks, Complex Systems 1:995-1019.
1. Saul and M. Jordan . (1994) Learning in Boltzmann Trees. Neural Comp o 6 :
1174-1184.
N. Sourlas. (1989) Spin Glass Models as Error Correcting Codes. Nature 339:
693-695 .
P. Stolorz. (1994) Links Between Dynamic Programming and Statistical Physics
for Heterogeneous Systems, JPL/Caltech preprint .
C. Williams and G. E. Hinton. (1990) Mean Field Networks That Learn To Discriminate Temporally Distorted Strings. Proc. Connectionist Models Summer School:
18-22.
| 966 |@word middle:3 polynomial:1 simulation:1 ajj:2 idl:1 initial:1 configuration:2 series:12 disparity:2 comparing:1 must:3 visible:11 partition:3 enables:1 mstep:1 designed:1 update:3 fewer:1 compo:1 constructed:1 combine:1 idr:1 baldi:2 notably:1 indeed:1 themselves:1 nor:1 mechanic:3 unfolded:1 provided:1 moreover:1 circuit:1 lame:1 ail:1 string:1 unified:1 firstorder:1 exactly:4 zl:1 unit:28 grant:1 consequence:1 acad:1 joining:3 might:1 exert:1 au:1 studied:1 equivalence:2 conversely:1 hmms:18 limited:1 range:4 implement:1 procedure:6 integrating:1 suggest:1 cannot:2 influence:1 equivalent:1 center:1 baum:4 williams:3 starting:1 duration:1 stolorz:2 welch:2 correcting:2 rule:21 adjusts:1 analogous:1 suppose:2 exact:3 programming:1 designing:1 decimation:4 roy:1 recognition:3 particularly:2 ze:1 approximated:1 observed:5 ackley:3 preprint:1 capture:1 connected:3 dempster:2 dynamic:1 trained:3 easily:1 aii:3 represented:3 alphabet:1 train:1 fast:1 shortcoming:4 effective:3 sejnowski:2 monte:1 prosodic:2 whose:2 statistic:2 laird:2 sequence:9 net:1 propose:1 interaction:1 product:4 neighboring:1 sourlas:2 combining:1 loop:4 flexibility:1 nonspectral:1 exploiting:1 juang:5 parent:1 transmission:1 oo:1 develop:2 coupling:4 ij:2 school:1 eq:3 soc:1 stochastic:1 human:1 eoo:2 offreedom:1 generalization:1 biological:1 hold:1 lawrence:5 bj:1 vary:2 estimation:5 proc:2 minimization:1 mit:2 clearly:1 rather:1 avoid:1 lifted:1 varying:2 jdr:2 derived:2 emission:4 modelling:1 likelihood:12 glass:2 bt:1 hidden:34 issue:1 constrained:1 special:5 mackay:4 jjl:1 field:6 f3:1 biology:1 represents:2 look:3 looped:4 warrant:1 future:1 connectionist:1 fundamentally:1 familiar:1 phase:3 technometrics:1 freedom:3 weakness:1 introduces:1 behind:1 chain:29 explosion:1 necessary:3 shorter:1 tfa:1 tree:3 incomplete:1 twostep:1 cda:1 modeling:8 maximization:4 resumed:1 successful:1 characterize:1 dependency:5 connect:3 amherst:1 probabilistic:3 physic:1 michael:5 quickly:2 na:1 connectivity:1 resort:1 toy:1 account:2 suggesting:1 jc:1 vi:1 onset:1 later:1 view:1 performed:1 closed:1 parallel:8 complicated:2 ii1:2 ass:1 il:5 spin:2 ni:1 who:1 efficiently:1 rabiner:2 resume:1 carlo:1 comp:1 submitted:1 ilil:1 energy:3 naturally:2 associated:1 couple:1 gain:2 massachusetts:1 organized:1 carefully:1 higher:1 cepstrum:2 done:5 though:2 anderson:2 correlation:9 until:1 hand:1 replacing:1 perhaps:1 usa:1 effect:1 contain:1 true:1 byrne:2 counterpart:1 analytically:1 alternating:2 laboratory:1 adjacent:1 complete:1 temperature:1 fj:1 percent:2 common:1 jl:4 analog:1 interpretation:1 occurred:1 cambridge:3 trivially:1 funded:1 belongs:1 bjj:1 reverse:1 certain:2 inequality:2 continue:1 success:1 caltech:1 additional:2 determine:1 maximize:2 signal:1 ii:1 full:1 corporate:1 characterized:1 long:9 mcclure:2 molecular:1 pitch:1 heterogeneous:1 essentially:1 expectation:3 represent:2 addition:1 whereas:1 annealing:1 extra:4 rest:1 exhibited:1 ascent:1 subject:1 thing:1 idd:1 jordan:6 call:3 iii:4 iterate:1 xj:1 fit:2 architecture:5 idea:1 knowing:1 whether:2 itzykson:1 speech:3 repeatedly:1 useful:1 clear:1 amount:1 statist:1 simplest:1 reduced:2 dangling:1 oversampling:1 nsf:1 designer:1 delta:2 per:1 jou:1 discrete:7 key:1 redundancy:1 neither:1 sum:4 inverse:1 powerful:1 distorted:1 family:2 ble:1 capturing:1 summer:1 occur:1 kronecker:1 idf:1 constraint:3 ri:1 generates:1 poor:1 across:1 slightly:1 em:4 making:1 computationally:1 equation:1 remains:1 tractable:10 end:2 operation:1 obey:1 spectral:2 enforce:1 jd:4 anv:1 exploit:1 establish:1 noticed:1 strategy:1 exhibit:1 gradient:2 separate:2 thank:1 simulated:1 atr:1 street:1 hmm:9 sci:2 link:1 luu:1 chauvin:2 reason:1 code:2 length:1 modeled:1 cnu:1 ratio:1 unfortunately:1 statement:1 relate:1 disparate:4 design:2 countable:1 motivates:1 boltzmann:44 lil:1 perform:1 markov:12 benchmark:1 finite:1 descent:1 t:1 situation:1 hinton:5 arbitrary:1 namely:2 mechanical:2 pair:3 required:1 connection:6 trans:1 address:4 below:1 built:3 lend:1 decidedly:1 solvable:1 technology:1 imply:1 temporally:1 mediated:1 coupled:3 naive:1 text:1 review:1 prior:1 epoch:4 acknowledgement:1 law:1 versus:4 validation:1 degree:6 rubin:2 decimating:3 normalizes:1 prone:1 periodicity:1 summary:1 last:1 free:1 aij:1 bias:3 zc:5 institute:1 saul:6 wide:1 peterson:2 regard:1 boundary:2 curve:1 transition:3 stand:1 jump:1 far:1 pruning:4 overfitting:3 summing:1 un:1 nature:1 transfer:1 learn:1 hunkapiller:2 complex:1 artificially:1 electric:1 arise:1 elaborate:1 slow:1 explicit:1 resistor:1 msec:2 wish:1 clamped:3 tied:3 tin:3 bij:7 cog:1 symbol:4 jpl:1 consist:1 nat:1 cartesian:4 led:1 likely:1 forming:1 ma:1 goal:1 viewed:1 typical:1 total:1 discriminate:1 e10:1 siemens:1 internal:1 tested:1 |
7,206 | 967 | An Integrated Architecture of Adaptive Neural Network
Control for Dynamic Systems
Robert L. Tokar2
Brian D.McVey 2
'Center for Nonlinear Studies, 2Applied Theoretical Physics Division
Los Alamos National Laboratory, Los Alamos, NM, 87545
Abstract
In this study, an integrated neural network control architecture for nonlinear dynamic systems is
presented. Most of the recent emphasis in the neural network control field has no error feedback as the
control input, which rises the lack of adaptation problem. The integrated architecture in this paper
combines feed forward control and error feedback adaptive control using neural networks. The paper
reveals the different internal functionality of these two kinds of neural network controllers for certain
input styles, e.g., state feedback and error feedback. With error feedback, neural network controllers
learn the slopes or the gains with respect to the error feedback, producing an error driven adaptive
control systems. The results demonstrate that the two kinds of control scheme can be combined to
realize their individual advantages. Testing with disturbances added to the plant shows good tracking
and adaptation with the integrated neural control architecture.
1 INTRODUCTION
Neural networks are used for control systems because of their capability to approximate nonlinear
system dynamics. Most neural network control architectures originate from work presented by
Narendra[I), Psaltis[2) and Lightbody[3) . In these architectures, an identification neural network is
trained to function as a model for the plant. Based on the neural network identification model, a neural
network controller is trained by backpropagating the error through the identification network. After
training, the identification network is replaced by the real plant. As is illustrated in Figure 1, the
controller receives external inputs as well as plant state feedback inputs. Training procedures are
employed such that the networks approximate feed forward control surfaces that are functions of
external inputs and state feedbacks of the plant (or the identification network during training).
It is worth noting that in this architecture, the error between the plant output and the desired output of
the reference model is not fed back to the controller, after the training phase. In other words, this error
information is ignored when the neural network applies its control. It is well known in control theory
that the error feedback plays a significant role in adaptation. Therefore, when model uncertainty or
noise/disturbances are present, a feed forward neural network controller with only state feedback will
not adaptively update the control signal. On line training for the neural controller has been proposed to
obtaip adaptive ability[I)[3). However, the stability for the on line training of the neural network
controller is unresolved[1][4] .
In this study, an additional nonlinear recurrent network is combined with the feed forward neural
network controller to form an adaptive controller. This added neural network uses feedback error
between the reference model output and the plant output as an input In addition, the system's external
Liu Ke, Robert L. Tokar, Brian D. McVey
1032
inputs and the plant states are also input to the feedback network. This architecture is used in the control
community, but not with neural network components. The approach differs from a conventional error
feedback controller, such as a gain scheduled PID controller, in that the neural network error feedback
controller implements a continuous nonlinear gain scheduled hypersurface, and after training, adaptive
model reference control for nonlinear dynamic systems is achieved without further parameter
computation. The approach is tested on well-known nonlinear control problems in the neural network
literature, and good results are obtained.
2 NEURAL NETWORK CONTROL
In this section, several different neural network control architectures are presented. In these structures,
identification neural networks, viewed as accurate models for real plants, are used.
2.1 NEURAL NETWORK FEED FORWARD CONTROL
The neural network controllers are trained by backpropagation of errors through a well trained neural
identification network. In this architecture, the state variable yet) of the system is sent back to the neural
network, and the external input x(t) also is input to the network. With these inputs, the neural network
estabJishes a feed forward mapping from the external input x(t) to the control signal u(t). This control
mapping is expressed as a function of the external input x(t) and the plant state yet) :
u(t)==j(x(t), yet?~
(1)
where x(t)=[x(t), x(t-l), .. J, andy(t)=[y(t), y(t-l), ..Y.
This neural network control architecture is denoted in this study as feed forward neural control even
though it includes state feedback. Neural control with error feedback is denoted as feedback neural
control.
x(t)
-..:...r-----~Ref.
x(t)
-~----~Ref.
Modelf-------,
Modell-----..,
e(t+ 1)
Control NN
y(t+ 1)
Figure I Neural Network Control Architecture.
ID NN represents the identification network.
Ref. Model means reference model, and NN
means neural network.
u(t) ,-------,
f----+-..
y(t+ 1)
Figure 2 Neural Network Feedback Control
Architecture
During the training phases, based on the assumption that the neural identification network provides a
model for the plant, the gradient information needed for error backpropagation is obtained by calculating
the Jacobian of the identification network. The following equation describes this process for the control
architecture shown in Figure I. If the cost function is defined as E, then the gradient of the cost function
with respect to weight w of the neural controller is
An Integrated Architecture of Adaptive Neural Network Control for Dynamic Systems
a E a E au (a E au aE Ja Yt-l
w + a u a Yt-l +a Yt-l
--a;-
a: = a; a
1033
(2)
where u is tbe control signal and YI-1 is tbe plant feedback state.
After tbe training stage, tbe neural network supplies a control law. Because neural networks have the
ability to approximate any arbitrary nonlinear functions[5], a feed forward neural network can build a
nonlinear controller, which is crucial to tbe use of tbe neural network in control engineering. Also, since
all tbe parameters of the neural network identification model and tbe neural network controller are
obtained from learning through samples, matbematically untraceable features of tbe plant can be
extracted from tbe samples and imbedded into tbe control system.
However, because tbe feed forward controller has no error feedback, tbe controller can not adapt to tbe
disturbances occurring in tbe plant or tbe reference model. This problem is of substantial importance in
tbe context of adaptive control. In tbe next subsection, error feedback between tbe reference models
and tbe plant outputs is introduced into neural network controllers for adaptation.
2.2 NEURAL ADAPTIVE CONTROL WITH ERROR FEEDBACK
It is known that feedback errors from the system are important for adaptation. Due to the flexibility of the
neural network architecture, the error between the reference model and the plant can be sent back to the
controller as an extra input. In such an architecture, neural networks become nonlinear gain scheduled
controllers with smooth continuous gains. Figure 2 shows the architecture for the feedback neural control.
With tbis architecture, tbe neural network control surface is not tbe fixed mapping from tbe x(t) to u(t)
for each state y(t), but instead it learns tbe slope or tbe gain referring to tbe feedback error e(t) for
control. This gain is a continuous nonlinear function of tbe external input x(t) and tbe state feedback
yet). Figure 3 shows tbe recurrent network architecture of tbe feedback neural controller. The output
node needs to be recurrent because tbe output witbout tbe recurrent link from tbe neural controller is
only a correction to tbe old control signal, and tbe new control signal should be tbe combination of old
control signal and tbe correction. The otber nodes of tbe network can be feed forward or recurrent. If
we denote tbe weight for tbe output node's recurrent link as w., tben tbe output from tbe recurrent link is
w.u(t-l). The following equation describes the feedback network.
u(t)
= wbu(t-I )+j(X(t), y(t), e(t?
(3)
where j(.) is a nonlinear function established by tbe network for which tbe recurrent link output is not
included and e(t)=[e(t), e(t-I), ... f
To compare tbe control gain expression with conventional control theory, consider tbe Taylor series
expansion of tbe network forward mappingj(.), equation (3) becomes
u(t) = w.u(t-l) +!'(x(t). yet?~ e(t)+ j"(x(t), yet?~ e 2(t)+...
where f'(x(t), y(t?=[ i1j(x(t), y(t), e(t?/ae(t), aJ!:x(t), y(t), e(t?Ii1e(t-I), ... ].
ignored and gO representsf'O, we get
u(t) =wbu(t-I)+ g(x(t), yet?~ e(t)
(4)
If high order terms are
(5)
1034
Liu Ke, Robert L. Tokar, Brian D. McVey
which is a gain scheduled controller and the gain is the function of external input x(/) and the plant state
It is clear that when w.=l.O, g(.) is a constant vector and e(/)=[e(t), e(t-l), e(t-2)]T, the feedback
neural network controller degenerates to a discrete PID controUer. Because the neural network can
approximate arbitrary nonlinear functions through learning, the neural network feedback controller can
generate a nonlinear continuous gain hypersurface.
y(/).
Ref. Model
Figure 3 Feedback Neural Network Controller
Figure 4 Integrated NN Control Architeture.
In the training process, error backpropagating through the identification network is used. The process is
similar to the training of a feed forward neural controller, but the resulting control surface is completely
different due to the different inputs. After training, the neural network is able to provide a nonlinear
control law, that is, the desired model following response can be obtained with fixed controller
parameters for nonlinear dynamic systems. Traditionally, the control of the nonlinear plant is derived
from continuous computing of the controller gains.
This feedback controller is error driven . As long as an error exists, the control signal is updated
according to the error and the gain. This kind of neural controller is an adaptive controller in principle.
2.3 INTEGRATED NEURAL NETWORK CONTROLLER
The characteristics of feed forward and error feedback neural control networks are described in the
previous subsections. In this section. the two controllers are combined. Figure 4 shows the architecture.
In this architecture, we include both feed forward and feedback neural network controllers. The control
signal is the combination from these two networks' outputs. In the training stage, it is our experience
that the feed forward network should be trained first. The feedback network is not included while
training the feed forward network. After training the feed forward controller, the error feedback network
is trained with the feed forward network, but the feed forward networks' weights are unchanged.
Backpropagating the error through the identification network is applied for the training of both
networks.
When training the feedback control network, the feed forward calculation is
=ujt)+u/b(t).
(6)
y(t+ 1) = P(x(t), y(t), u(t?,
(7)
u(t)
where uj/) is the output from the feed forward controller network and u,..(t) is the output from the
feedback controller network, P(.) is the identification mapping.
1035
An Integrated Architecture of Adaptive Neural Network Control for Dynamic Systems
3 CONTROL ON EXAMPLE PROBLEMS
In this section, the control architecture described above is applied to a well-known problem from the
literature[I). The plants and the reference model of the sample problems are described by difference
equations
y(t + 1)
plant:
=
yet)
2
+ (u(t) -1.O)u(t)(u(t) + 1.0)
1.0+ Y (t)
(II)
y(t + 1) = 0.6y(t) + u(t)
reference model:
(12)
This is a nonlinear time varying dynamic system with no analytical inverse.
3.1 FEED FORWARD CONTROL
A feed forward neural network is trained to control the system to follow the reference model. The plant
state yet) and external inputx(t) are fed to the controller. During the training, the x(t) is randomly
generated. After training, the controller generates a control signal u(t) such that the plant can follow the
reference model output. Figure 5 shows the testing result of the reference model output and the
controlled plant output. The input function is x(t)=sin(21ttf25)+sin(21tt/1O). The controller network
architecture is (2, 20, 1).
4
2
OJ
0
c
Q)
2
2
?
(j
10-
OJ
~
'"
1J
::J
0
'"
0
c
... 1
~
c
0
0
()
0
u
-2
20
40
60
80
~0
-1
...2
100
Figure 5 Tracking Result From the Feed Forward NN.
Output of reference (solid line) and plant (dash line).
Figure 6 Feed Forward Control Surface
The output surface of the controller network is shown in Figure 6. By examining the controller output
surface, we can see that the neural network builds a feed forward mapping from x(t) to u(t). This feed
forward mapping is also a function of the plant state yet). Under each state, the neural network
controller accepts input x(t) to produce control signal u(t) such that the plant follows the reference model
reasonably well. In Figure 6, the x axis is the external input x(t) and the y axis is the plant feedback
output yet) . The z axis represents the control surface.
The feed forward controller laCks the ability to adapt to plant uncertainty, noise or changes in the
reference model. As an example, we apply this feed forward controller to the disturbed plant with a bias
0.5 added to the original plant. The tracking result is shown in Figure 7. With this slight bias, the plant
does not follow the reference model. Clearly, the feed forward controller has no adaptive ability to this
model bias.
1036
Liu Ke, Robert L. Tokar, Brian D. McVey
3.2 FEEDBACK CONTROL
FtrSt, we compare the neural network feedback controller with fixed gain PID controllers. For many
nonlinear systems, the fixed gain PID controllers will give poor tracking and continuous adaptation of
the controller parameters is needed. The neural network approach offers an alternative control approach
for nonlinear systems. Through the training, control gains, imbedded in the neural network, are
established as a continuous function of system external inputs x(t) and plant states yet).
The sample problem in the above section is now employed to describe how the neural network creates a
nonlinear control gain surface with error feedback and additional inputs. First, we show one simple case
of neural adaptive feedback controller. This controller can only adapt to the system nonlinearity with a
fixed linear input pattern. The reason to show this simple adaptation case first is that its control gain
surface can be illustrated graphically.
Figure 8 illustrates, for the system in equations (11) and (12) that a fixed gain PI controller fails to track
the reference model, for even one fixed linear input pattern x(t)=0.2t-2.5, because the plant nonlinearity.
Figure 9 illustrates the result from a recurrent neural network with feedback error e(t) and x(t) as inputs.
The neural network is trained by backpropagation error through the identification network. Compared to
the flXed gain PI controller, the neural network improves the tracking ability significantly.
.,
.,.c
.,
4
>.
0
6
'!!
-.::I
C
0
>.
,
OJ
0
<J
C
~
2
.2
~
I
I
3
0
>.
'0
c
-3
0
-2
>.
20
40
60
80
100
t
Figure 7 Tracking Result for Shifted Plant, plant
output (dash line) and reference output (solid line).
o
5
10
15
20
25
30
35
t
Figure 8 Reference Model Output (solid line)
and PID Controlled Plant Output (dashed line)
The control surface of the updating output fl.) is shown in Figure 10, which is the output from the neural
network controller without recurrent link (see equation (3?. We plot the surface of the updating output
from the controller with respect to input x(t) and error feed back input e(t). The gain of the controller is
equivalent to the updating output from the network when error=l.O. As shown in the figure, the gain in
the neighborhood about x(t)=O changes largely according to the direction of changes in the plant in the
corresponding region. The updating surface for a PID controller is a plane. The neural network
implements a nonlinear continuous control gain surface.
For a more complicated case, we addx(t-I) as another input to the neural network as well as e(t-l), and
train by error backpropagation through the identification network. These two inputs, x(t) and x(t-I) add
difference information to the network. The network can adapt to not only different operating regions
indicated by x(t), but also different input patterns. Figure 11 shows the tracking results with two
different input patterns. In Figure II (a), input pattern is x(t)=4.0sin(tI4.0). In Figure 11 (b) input
pattern is x(t)=sin(21t1!25)+sin(21t111O) .
An Integrated Architecture of Adaptive Neural Network Control for Dynamic Systems
1037
6
Q)
u
c
~
2
~
3
0
,
,
>.
-0.
c
0
>.
I
I
-3
-6
o
5
10
15
20
25
30
35
t
Figure 9 Reference Model Output (solid line) and
Neural Network Controled Output (dashed line)
Figure 10 Feedback Neural Controller Updating Surface
.,u
OJ
u
c
~
~
~
c
~
OJ
l'
0
-2
"c:0 - 4
>. -6
-8
-10
>.
>.
"
c
5
4
3
2
1
0
-1
-2
0
-3
>.
- 4
-5
-6
20
0
40
60
80
10C
(a)
0
20
40
60
80
100
(b)
Figure 11 Output of the Reference Model (solid line) and the Plant (dash line)
3.3 INTEGRA TED NEURAL CONTROLLER
As shown in the above section, when only error feedback neural controller is used, the control result is
not very accurate. Now we combine feed forward and feedback control to realize good tracking and
adaptation. Figure 12 shows the control result from the integrated controller when the plant is shifted
O.S. Compared to only feed forward control (Figure 7), the integrated controller has much better
adaptation to the shifted plant.
When the plant changes, adding an extra feed back controller can avoid on-line training of feed forward
network which may induce potential instability, and the adaptation is achieved. The output from the
feedback network controller is driven by the error between the reference model and the plant.
4 DISCUSSIONS
We have emphasized in the above sections that a feed forward controller with only state feedback does
not adapt when model uncertainties or noise/disturbance are present. The presence of a feed back
controller can make the on line training of the feed forward network unnecessary, thus avoiding
potential instability. The main reason for the instability of on-line training is the incompleteness of
sample sets, which is referred to as a lack of persistent excitation in control theory[6]. First, it leads to
an inaccurate identification network. Training with this network can result in an unstable controller.
Second, it makes the training of controller away from global representation . With an error feedback
adaptive network, the output from the feedback network controller is driven by the error between the
reference model and the plant. In the simplest case when all the activity functions are linear and only
the feedback errors are inputs, this kind of neural network is equivalent to a PID controller. However,
Liu Ke, Robert L Tokar, Brian D. McVey
1038
beyond the scope of PID controllers, the neural networks are capable to approximating nonlinear time
variant control gain surfaces corresponding to different operating regions. Also, unlike a PID controller,
the coefficients for the neural adaptive controller are obtained through a training procedure.
4
Q)
u
c
1:'
Q)
0;
...
>,
"0
C
0
>,
o
20
40
60
80
100
Figure 12 Integrated Network Controller Tracking Result for Shifted Plant.
Plant Output (dash line) and Reference Output (solid line).
The error feedback network behaves as a gain scheduling controller. It has rise time, overshoot
consideration and delay problem. Feed forward control can compensate for these problems to some
degree. For example, the feed forward network can perform a nonlinear mapping with designed time
delay. Therefore with the feed forward network, the delay problem maybe overcame significantly. Also
the feed forward controller can help to reduce rise time compare to use only feedback controller.
With the feed forward network, the feedback network controller can have much smaller gains compared
to using a feedback network alone. This increases the noise rejection ability. Also this reduces the
overshoot as well as settle time.
The neural network control architecture offers an alternative to the conventional approach. It gives a
generic model for the broadest class of systems considered in control theory . However this model needs
to be configured depending on the details of the control problem. With different inputs, the neural
network controllers establish different internal hyperstates. When plant states are fed back to the
network, a feed forward mapping is established as a function of the plant states by the neural network
controller. When the errors between the reference model and the plant are used as the error feedback
inputs to a dynamic neural network controller, the network functions as an associative memory
nonlinear gain scheduled controller. The above two kinds of neural controller can be combined and
complemented to achieve accurate tracking and adaptation.
References
[1] Kumpati S. Narendra and Kannan Parthasarathy. "Gradient Methods for the Optimization of DynamiCal
Systems Containing Neural Networks," IEEE Trans. Neural Networks. vol. 2. pp252-262 Mar. 1991
[2] Psaltis. D .? Sideris. A. and Yamamura. A., "Neural controllers." Proc. of 1st International Conference on
Neural Networks. Vol. 4. pp551-558. San Diego. CA. 1987
[3) G. lightbOdy. Q. H. Wu and G. W. Irwin. "Control applications for feed forward networks." Chapter 4.
Neural Networks for Control and Systems. Edited by K.warwich, G. W. Irwin and K. J. Hunt 1992
[4) R. Abikowski and P. 1. Gawthrop. "A survey of neural networks for control" Chapter 3. NeUral Networks
for Control and Systems. ISBN 0-86341-279-3. Edited by K.warwich. G. W. Irwin and K. 1. Hunt 1992
[5] John Hertz. Anders Krogh and Richard G. Palmer. "Introduction to the Theory of Neural Computation."
[6J Thomas Miller. RiChard S. Sutton and Paul 1. Werbos. "Neural Networks for Control"
| 967 |@word hyperstates:1 solid:6 liu:4 series:1 yet:12 john:1 realize:2 plot:1 designed:1 update:1 alone:1 plane:1 provides:1 node:3 become:1 supply:1 persistent:1 combine:2 becomes:1 kind:5 control:86 producing:1 t1:1 engineering:1 sutton:1 id:1 emphasis:1 au:2 hunt:2 palmer:1 testing:2 implement:2 differs:1 backpropagation:4 witbout:1 procedure:2 significantly:2 word:1 induce:1 get:1 scheduling:1 context:1 instability:3 disturbed:1 conventional:3 equivalent:2 center:1 yt:3 go:1 graphically:1 survey:1 ke:4 stability:1 traditionally:1 updated:1 diego:1 play:1 us:1 updating:5 werbos:1 role:1 region:3 edited:2 substantial:1 dynamic:10 trained:8 overshoot:2 creates:1 division:1 completely:1 chapter:2 train:1 describe:1 neighborhood:1 ability:6 otber:1 associative:1 advantage:1 analytical:1 isbn:1 unresolved:1 adaptation:11 flexibility:1 degenerate:1 achieve:1 los:2 produce:1 help:1 depending:1 recurrent:10 krogh:1 direction:1 functionality:1 settle:1 ja:1 brian:5 correction:2 considered:1 mapping:8 scope:1 narendra:2 proc:1 psaltis:2 clearly:1 avoid:1 varying:1 derived:1 anders:1 nn:5 inaccurate:1 integrated:12 denoted:2 field:1 ted:1 represents:2 richard:2 randomly:1 national:1 individual:1 replaced:1 phase:2 flxed:1 accurate:3 andy:1 capable:1 experience:1 old:2 taylor:1 desired:2 theoretical:1 cost:2 alamo:2 delay:3 examining:1 yamamura:1 combined:4 adaptively:1 referring:1 st:1 international:1 physic:1 nm:1 containing:1 external:11 style:1 potential:2 includes:1 coefficient:1 configured:1 capability:1 complicated:1 slope:2 characteristic:1 largely:1 miller:1 identification:17 worth:1 broadest:1 gain:27 subsection:2 improves:1 back:7 feed:44 follow:3 response:1 though:1 mar:1 stage:2 receives:1 nonlinear:25 lack:3 aj:1 indicated:1 scheduled:5 laboratory:1 illustrated:2 sin:5 during:3 backpropagating:3 excitation:1 tt:1 demonstrate:1 consideration:1 behaves:1 slight:1 significant:1 nonlinearity:2 modell:1 surface:15 operating:2 add:1 recent:1 driven:4 certain:1 sideris:1 yi:1 additional:2 employed:2 integra:1 signal:10 ii:2 dashed:2 reduces:1 smooth:1 adapt:5 calculation:1 offer:2 long:1 compensate:1 controlled:2 variant:1 controller:88 ae:2 achieved:2 addition:1 crucial:1 extra:2 unlike:1 sent:2 noting:1 presence:1 architecture:27 reduce:1 expression:1 ignored:2 clear:1 maybe:1 simplest:1 generate:1 shifted:4 track:1 discrete:1 controled:1 vol:2 tbe:47 inverse:1 uncertainty:3 wu:1 incompleteness:1 fl:1 overcame:1 dash:4 activity:1 generates:1 according:2 combination:2 poor:1 hertz:1 describes:2 smaller:1 i1j:1 pid:9 equation:6 needed:2 fed:3 apply:1 away:1 generic:1 alternative:2 original:1 thomas:1 include:1 calculating:1 build:2 uj:1 approximating:1 establish:1 unchanged:1 added:3 imbedded:2 gradient:3 link:5 originate:1 unstable:1 reason:2 kannan:1 ftrst:1 robert:5 rise:3 perform:1 arbitrary:2 community:1 introduced:1 accepts:1 established:3 trans:1 able:1 beyond:1 dynamical:1 pattern:6 ujt:1 oj:5 memory:1 disturbance:4 scheme:1 axis:3 parthasarathy:1 literature:2 law:2 plant:48 degree:1 principle:1 pi:2 bias:3 feedback:57 forward:42 adaptive:16 san:1 hypersurface:2 approximate:4 global:1 reveals:1 unnecessary:1 continuous:8 learn:1 reasonably:1 ca:1 expansion:1 main:1 noise:4 paul:1 ref:4 referred:1 fails:1 jacobian:1 learns:1 emphasized:1 exists:1 adding:1 importance:1 illustrates:2 occurring:1 rejection:1 expressed:1 tracking:10 applies:1 extracted:1 complemented:1 viewed:1 change:4 included:2 internal:2 irwin:3 tested:1 avoiding:1 |
7,207 | 968 | Capacity and Information Efficiency of a
Brain-like Associative Net
Bruce Graham and David Willshaw
Centre for Cognitive Science, University of Edinburgh
2 Buccleuch Place, Edinburgh, EH8 9LW, UK
Email: bruce@cns.ed.ac.uk&david@cns.ed.ac.uk
Abstract
We have determined the capacity and information efficiency of an
associative net configured in a brain-like way with partial connectivity and noisy input cues. Recall theory was used to calculate
the capacity when pattern recall is achieved using a winners-takeall strategy. Transforming the dendritic sum according to input
activity and unit usage can greatly increase the capacity of the
associative net under these conditions. For moderately sparse patterns, maximum information efficiency is achieved with very low
connectivity levels (~ 10%). This corresponds to the level of connectivity commonly seen in the brain and invites speculation that
the brain is connected in the most information efficient way.
1
INTRODUCTION
Standard network associative memories become more plausible as models of associative memory in the brain if they incorporate (1) partial connectivity, (2) sparse
activity and (3) recall from noisy cues. In this paper we consider the capacity of
a binary associative net (Willshaw, Buneman, & Longuet-Higgins, 1969; Willshaw,
1971; Buckingham, 1991) containing these features. While the associative net is
a very simple model of associative memory, its behaviour as a storage device is
not trivial and yet it is tractable to theoretical analysis. We are able to calculate
514
Bruce Graham, David Willshaw
the capacity of the net in different configurations and with different pattern recall
strategies. Here we consider the capacity as a function of connectivity level when
winners-take-all recall is used.
The associative net is an heteroassociative memory in which pairs of binary patterns are stored by altering the connection weights between input and output units
via a Hebbian learning rule. After pattern storage, an output pattern is recalled
by presenting a previously stored input pattern on the input units. Which output
units become active during recall is determined by applying a threshold of activation to measurements that each output unit makes of the input cue pattern. The
most commonly used measurement is the weighted sum of the inputs, or dendritic
sum. Amongst the simpler thresholding strategies is the winners-take-all (WTA)
approach, which chooses the required number of output units with the highest dendritic sums to be active. This works well when the net is fully connected (each input
unit is connected to every output unit), and input cues are noise-free. However,
recall performance deteriorates rapidly if the net is partially connected (each input
unit is connected to only some of the output units) and cues are noisy.
Marr (1971) recognised that when an associative net is only partially connected,
another useful measurement for threshold setting is the total input activity (sum of
the inputs, regardless of the connection weights). The ratio of the dendritic sum
to the input activity can be a better discriminator of which output units should be
active than the dendritic sum alone. Buckingham and Willshaw (1993) showed that
differences in unit usage (the number of patterns in which an output unit is active
during storage) causes variations in the dendritic sums that makes accurate recall
difficult when the input cues are noisy. They incorporated both input activity and
unit usage measurements into a recall strategy that minimised the number of errors
in the output pattern by setting the activity threshold on a unit by unit basis. This
is a rather more complex threshold setting strategy than a simple winners-take-all.
We have previously demonstrated via computer simulations (Graham & Wills haw ,
1994) that the WTA threshold strategy can achieve the same recall performance
as this minimisation approach if the dendritic sums are transformed by certain
functions of the input activity and unit usage before a threshold is applied. Here
we calculate the capacity of the associative net when WTA recall is used with three
different functions of the dendritic sums: (1) pure dendritic sums, (2) modified by
input activity and (3) modified by input activity and unit usage. The results show
that up to four times the capacity can be obtained by transforming the dendritic
sums by a function of both input activity and unit usage. This increase in capacity
was obtained without a loss of information efficiency. For the moderately sparse
patterns used, WTA recall is most information efficient at low levels of connectivity
(~ 10%), as is the minimisation approach to threshold setting (Buckingham, 1991).
This connectivity range is similar to that commonly seen in the brain.
Capacity and Infonnation Efficiency of a Brain-Like Associative Net
2
515
NOTATION AND OPERATION
The associative net consists of N B binary output units each connected to a proportion Z of the N A binary input units. Pairs of binary patterns are stored in the net.
Input and output patterns contain MA and MB active units, respectively (activity
level a = M / N ? 1). All connection weights start at zero. On presentation to
the net of a pattern pair for storage, the connection weight between an active input
unit and an active output unit is set to 1. During recall an input cue pattern is
presented on the input units. The input cue is a noisy version of a previously stored
input pattern in which a fraction, s, of the MA active units do not come from the
stored pattern. A thresholding strategy is applied to the output units to determine
which of them should be active. Those that should be active in response to the
input cue will be called high units, and those that should be inactive will be called
low units. We consider winners-take-all (WTA) thresholding strategies that choose
to be active the MB output units with the highest values of three functions of the
dendritic sum, d, the input activity, a, and the unit usage, r. These functions are
listed in Table 1. The normalised strategy deals with partial connectivity. The
transformed strategy reduces variations in the dendritic sums due to differences in
unit usage. This function minimises the variance of the low unit dendritic sums
with respect to the unit usage (Graham & Willshaw, 1994).
Table 1: WTA Strategies
WTA Strategy
Basic
Normalised
Transformed
3
FUnction
d
d' = d/a
d* = 1 - (1 - d/a)l/r
RECALL THEORY
The capacity of the associative net is defined to be the number of pattern pairs that
can be stored before there is one bit in error in a recalled output pattern. This
cannot be calculated analytically for the net configuration under study. However, it
can be determined numerically for the WTA recall strategy by calculating the recall
response for different numbers of stored patterns, R, until the minimum value of R
is found for which a recall error occurs. The WTA recall response can be calculated
theoretically using expressions for the distributions of the dendritic sums of low and
high output units. The probability that the dendritic sum of a low or high output
unit should have a particular value x is, respectively (Buckingham & Willshaw,
1993; Buckingham, 1991)
P(d,
= x) =
t. (~ )
"'8(1- "B)R-,
( "!,A ) (Zp[rJ)"(I- Zp[rJ)MA-"
(I)
516
Bruce Graham. David Willshaw
P(dh = x)
= ~ ( ~ ) a;'(l-aB)R-" ( ~A
) (ZI'[r +1])"(1- ZI'[r+1])MA-"
(2)
where p[r] and /L[r] are the probabilities that an arbitrarily selected active input is
on a connection with weight 1. For a low unit, p[r] = 1- (1- OA)r. For a high unit
a good approximation for /L is /L[r + 1] ~ g + sp[r] = 1 - s(l - OAY where g and
s are the probabilities that a particular active input in the cue pattern is genuine
(belongs to the stored pattern) or spurious, respectively (g + s = 1) (Buckingham
& Willshaw, 1993). The basic WTA response is calculated using these distributions
by finding the threshold, T, that gives
(3)
The number of false positive and false negative errors of the response is given by
(4)
The actual distributions of the normalised dendritic sums are the distributions of
dja. For the purposes of calculating the normalised WTA response, it is possible to use the basic distributions for the situation where every unit has the mean
input activity, am = MAZ. In this case the low and high unit distributions are
approximately
P(til
P(d'h
= x) =
t. (~ )
,,;'(1- "B)R-" ( "; ) (p[r])"(l- p[r])"m-"
= x) = ~ ( ~ ) aB(1- "B )R-" (
(5)
a; ) (I'[r + 1])" (l-l'[r +1])"m -" (6)
Due to the nonlinear transformation used, it is not possible to calculate the transformed distributions as simple sums of binomials, so the following approach is used
to generate the transformed WTA response. For a given transformed threshold, T*,
and for each possible value of unit usage, r, an equivalent normalised threshold is
calculated via
T'[r] = am (l - (1 - T*Y)
(7)
The transformed cumulative probabilities can then be calculated from the normalised distributions:
P(dj :2: TO) =
P( di. :2: T')
t. (~ )
aB(1 - "B )R-" P( til :2: T'[r])
= ~ ( ~ ) ,,;'(1 -
"B )R-" P(d' :2: T'[r + 1])
(8)
(9)
The normalised and transformed WTA responses are calculated in the same manner
as the basic response, using the appropriate probability distributions.
Capacity and Information Efficiency of a Brain-Like Associative Net
(a) 0% noise
5000
g
~
?u
--B
- - N, T
(b) 40% noise
...
.....................
~
.-~
...-.,
3000
--B
?-N
2500
---T
./"
4000
517
2000
3000
1500
ca
Co
ca 2000
U
1000
1000
500
O~---...L
20
40
60
80
Connectivity (%)
100
0
___--'---'-.....L-....o...-..I.-..........J
20
40
60
80
100
Connectivity (%)
Figure 1: Capacity Versus Connectivity
4
RESULTS
Extensive simulations were previously carried out of WTA recall from a large
associative net with the following specifications (Graham & Willshaw, 1994):
NA = 48000, MA = 1440, NB = 6144, MB = 180. Agreement between the simulations and the theoretical recall described above is extremely good, indicating that
the approximations used in the theory are valid. Here we use the theoretical recall
to calculate capacity results for this large associative net that are not easily obtained via simulations. All the results shown have been generated using the theory
described in the previous section.
Figure 1 shows the capacity as a function of connectivity for the different WTA
strategies when there is no noise in the input cue, or 40% noise in the cue (legend: B
= basic WTA, N = normalised WTA, T = transformed WTA; for clarity, individual
data points are omitted). With no noise in the cue the normalised and transformed
methods perform identically, so only the normalised results are shown. Figure l(a)
highlights the effectiveness of normalising the dendritic sums against input activity
when the net is partially connected. Figure 1(b) shows the effect of noise on capacity.
The capacity of each recall strategy at a given connectivity level is much reduced
compared to the noise-free case. However, for connectivities greater than 10% the
capacity of the transformed WTA is now much greater than either the normalised
or basic WTA.
The relative capacities of the different strategies are shown in Figure 2 (legend:
NIB = ratio of normalised to basic capacity, T I B = ratio of transformed to basic,
TIN = ratio of transformed to normalised). In the noise-free case (Figure 2(a)), at
low levels of connectivity the relative capacity is distorted because the basic capacity
518
Bruce Graham, David Willshaw
drops to near zero, so that even low normalised capacities are relatively very large.
For most connectivity levels (10-90%) the normalised WTA provides 2-4 times the
capacity of the basic WTA. In the noisy case (Figure 2(b)), the normalised capacity
is only up to 1.5 times the basic capacity over this range of connectivities. The
transformed WTA, however, provides 3 to nearly 4 times the basic capacity and 2.5
to nearly 3 times the normalised capacity for connectivities greater than 10%.
The capacities can be interpreted in information theoretic terms by considering the
information efficiency of the net. This is the ratio of the amount of information
that can be retrieved from the net to the number of bits of storage available and
is given by "'0 = Ro10jZNANB' where Ro is the capacity, 10 is the amount of
information contained in an output pattern and Z NANB is the number of weights,
or bits of storage required (Willshaw et al., 1969; Buckingham & Willshaw, 1992).
Information efficiency as a function of connectivity is shown in Figure 3. There is
a distinct peak in information efficiency for each of the recall strategies at some
low level of connectivity. The peak information efficiencies and the efficiencies at
full connectivity are summarised in Table 2. The greatest contrast between full
and partial connectivity is seen with the normalised WTA and noise-free cues. At
1% connectivity the normalised WTA is nearly 14 times more efficient than at full
connectivity. In absolute terms, however, the normalised capacity is only 694 at
1% connectivity, compared with 5122 at full connectivity. The peak efficiency of
53% obtained by the normalised WTA is approaching the theoretically approximate
maximum of 69% for a fully connected net (Willshaw et al., 1969).
5
DISCUSSION
Previous simulations (Graham & Willshaw, 1994) have shown that, when the input
cues are noisy, the recall performance of the winners-take-all thresholding strategy
applied to the partially connected associative net is greatly improved if the dendritic sums of the output units are transformed by functions of input activity and
unit usage. We have confirmed and extended these results here by calculating the
theoretical capacity of the associative net as a function of connectivity.
For the moderately sparse patterns used, all of the recall strategies are most information efficient at very low levels of connectivity (~ 10%). However, the optimum
connectivity level is dependent on the pattern coding rate. Extending the analysis
of Willshaw et al. (1969) to a partially connected net using normalised WTA recall
yields that maximum information efficiency is obtained when ZMA = log2(NB).
So for input coding rates higher than log2(NB), a partially connected net is most
information efficient. For the input coding rate used here, this relationship gives an
optimum connectivity level of 0.87%, very close to the 1% obtained from the recall
theory.
Comparing the peak efficiencies across the different strategies for the noisy cue
case, the normalised WTA is about twice as efficient as the basic WTA, while the
transformed WTA is three times as efficient. This comparison does not include the
Capacity and Information Efficiency of a Brain-Like Associative Net
(a) 0% noise
(b) 40% noise
12
4
- - NIB, TIB
10
.~
0
III
a.
I
~
;;
,
3
(
8
III
Q)
~------
6
4
a:
0
40
60
20
80
Connectivity (%)
0
,
I I
\1 '
......
- ..............
1;?-
,,
- -___
I, - - NIB
II
1
-_
' --- T/B
I
TIN
\
100
..... , ,
,..- ...... -------- -~
}{' ...
2
.,
2
""
I
III
u
519
0
20
40
60
80
Connectivity (%)
0
100
Figure 2: Relative Capacity Versus Connectivity
(a) 0% noise
(b) 40% noise
60
-
6
~ 50
g
Q)
40
ffic:
30
.(3
ia 20
E
....
0
1: 10
o
4
\
.Q
~
r-o
3
2
----"::7
20
40
60
80
Connectivity (%)
,,,'
5
--B
._._- N, T
>.
",
100
0
--B
--N
---T
'\ '\
'\
'\
,
f.l
.\
I
'" "
.,
;'\
., ' ...'--
"
............
- ...
---------
L-'--'........- - - L _ - - L _ - - - L . - - ' - - - I
0
20
40
60
80
Connectivity (%)
Figure 3: Information Efficiency Versus Connectivity
Table 2: Information Efficiency
WTA
Strategy
Basic
Normalised
Transformed
at
Peak
(%)
6.1
53.3
53.3
TJo
0% Noise
Z at
TJo at
Peak Z=1
(%)
(%)
4
3.9
1
3.9
1
3.9
at
Peak
T]o
40% Noise
Z at
TJo at
Peak Z=1
(%)
(%)
(%)
2.0
3.6
5.7
7
5
10
0.8
0.8
2.3
100
520
Bruce Graham, David Willshaw
cost of storing input activity and unit usage information. If one bit of storage per
connection is required for the input activity, and another bit for the unit usage, then
the information efficiency of the normalised WTA is halved, and the information
efficiency of the transformed WTA is reduced by two thirds. This results in all the
strategies having about the same peak efficiency. However, the absolute capacities
of the different strategies at their peak efficiencies are 183,237 and 741 for the basic,
normalised and transformed WTA, respectively. So, at the same level of efficiency,
the transformed WTA delivers four times the capacity of the basic WTA.
In conclusion, numerical calculations of the capacity of the associative net show that
it is most information efficient at a very low level of connectivity when moderately
sparse patterns are stored. Including input activity and unit usage information into
the recall calculations results in a four-fold increase in storage capacity without loss
of efficiency.
Acknowledgements
To the Medical Research Council for financial support under Programme Grant PG
9119632
References
Buckingham, J., & Willshaw, D. (1992). Performance characteristics of the associative net. Network, 8, 407-414.
Buckingham, J., & Willshaw, D. (1993). On setting unit thresholds in an incompletely connected associative net. Network, 4, 441-459.
Buckingham, J. (1991). Delicate nets, faint recollections: a study of partially connected associative network memories. Ph.D. thesis, University of Edinburgh.
Graham, B., & Willshaw, D. (1994). Improving recall from an associative memory.
Bioi. Cybem., in press.
Marr, D. (1971). Simple memory: a theory for archicortex. Phil. Trans. Roy. Soc.
Lond. B, 262, 23-81.
Shepherd, G. (Ed.). (1990). The Synaptic Organization of the Brain (Third edition).
Oxford University Press, New York, Oxford.
Willshaw, D. (1971). Models of distributed associative memory. Ph.D. thesis, University of Edinburgh.
Willshaw, D., Buneman, 0., & Longuet-Higgins, H. (1969). Non-holographic associative memory. Nature, 222, 960-962.
| 968 |@word soc:1 effect:1 contain:1 maz:1 version:1 come:1 proportion:1 analytically:1 occurs:1 strategy:23 simulation:5 deal:1 heteroassociative:1 pg:1 during:3 amongst:1 incompletely:1 behaviour:1 capacity:39 oa:1 configuration:2 recollection:1 presenting:1 recognised:1 dendritic:18 theoretic:1 trivial:1 delivers:1 comparing:1 relationship:1 activation:1 yet:1 buckingham:10 archicortex:1 ratio:5 difficult:1 numerical:1 dja:1 negative:1 omitted:1 drop:1 purpose:1 winner:6 alone:1 cue:16 selected:1 device:1 numerically:1 infonnation:1 measurement:4 perform:1 council:1 weighted:1 normalising:1 situation:1 provides:2 centre:1 extended:1 incorporated:1 dj:1 modified:2 simpler:1 rather:1 specification:1 become:2 minimisation:2 halved:1 david:6 consists:1 showed:1 retrieved:1 pair:4 belongs:1 required:3 speculation:1 connection:6 manner:1 certain:1 theoretically:2 greatly:2 contrast:1 binary:5 arbitrarily:1 am:2 recalled:2 eh8:1 trans:1 brain:10 dependent:1 seen:3 minimum:1 greater:3 able:1 pattern:26 actual:1 spurious:1 determine:1 considering:1 transformed:20 ii:1 notation:1 full:4 memory:9 rj:2 reduces:1 including:1 hebbian:1 greatest:1 ia:1 interpreted:1 calculation:2 genuine:1 finding:1 transformation:1 having:1 every:2 buneman:2 basic:16 nearly:3 carried:1 willshaw:22 ro:1 uk:3 unit:47 medical:1 grant:1 achieved:2 discriminator:1 acknowledgement:1 before:2 positive:1 individual:1 relative:3 fully:2 loss:2 highlight:1 cns:2 oxford:2 delicate:1 ab:3 extensive:1 versus:3 organization:1 approximately:1 shepherd:1 twice:1 legend:2 effectiveness:1 thresholding:4 co:1 tjo:3 near:1 storing:1 range:2 iii:3 identically:1 accurate:1 zi:2 free:4 approaching:1 partial:4 normalised:26 inactive:1 absolute:2 expression:1 sparse:5 edinburgh:4 distributed:1 theoretical:4 calculated:6 valid:1 cumulative:1 commonly:3 cannot:1 close:1 altering:1 york:1 cause:1 storage:8 nb:3 applying:1 cost:1 useful:1 programme:1 listed:1 equivalent:1 approximate:1 demonstrated:1 phil:1 holographic:1 amount:2 regardless:1 ph:2 active:13 cybem:1 stored:9 reduced:2 generate:1 pure:1 chooses:1 rule:1 higgins:2 peak:10 marr:2 deteriorates:1 per:1 financial:1 table:4 nature:1 summarised:1 minimised:1 variation:2 longuet:2 ca:2 improving:1 na:1 connectivity:37 thesis:2 four:3 threshold:11 complex:1 containing:1 choose:1 clarity:1 agreement:1 sp:1 roy:1 cognitive:1 noise:16 til:2 edition:1 fraction:1 sum:21 tib:1 coding:3 distorted:1 calculate:5 place:1 configured:1 connected:14 highest:2 graham:10 transforming:2 bit:5 start:1 moderately:4 lw:1 third:2 tin:2 fold:1 bruce:6 activity:18 variance:1 characteristic:1 nib:3 efficiency:23 yield:1 basis:1 faint:1 easily:1 false:2 extremely:1 lond:1 confirmed:1 distinct:1 relatively:1 according:1 ed:3 email:1 synaptic:1 across:1 against:1 plausible:1 wta:36 contained:1 buccleuch:1 partially:7 di:1 noisy:8 corresponds:1 associative:27 dh:1 ma:5 recall:29 previously:4 net:33 bioi:1 presentation:1 mb:3 tractable:1 haw:1 rapidly:1 higher:1 operation:1 available:1 determined:3 response:9 achieve:1 improved:1 takeall:1 appropriate:1 total:1 called:2 until:1 optimum:2 zp:2 extending:1 invite:1 binomial:1 nonlinear:1 include:1 indicating:1 support:1 log2:2 ac:2 minimises:1 incorporate:1 calculating:3 usage:14 |
7,208 | 969 | Adaptive Elastic Input Field for
Recognition Improvement
Minoru Asogawa
C&C Research Laboratories, NEe
Miyamae, Miyazaki, Kawasaki Kanagawa 213 Japan
asogawa~csl.cl.nec.co.jp
Abstract
For machines to perform classification tasks, such as speech and
character recognition, appropriately handling deformed patterns
is a key to achieving high performance. The authors presents a
new type of classification system, an Adaptive Input Field Neural Network (AIFNN), which includes a simple pre-trained neural
network and an elastic input field attached to an input layer. By
using an iterative method, AIFNN can determine an optimal affine
translation for an elastic input field to compensate for the original
deformations. The convergence of the AIFNN algorithm is shown.
AIFNN is applied for handwritten numerals recognition. Consequently, 10.83% of originally misclassified patterns are correctly
categorized and total performance is improved, without modifying
the neural network.
1
Introduction
For machines to accomplish classification tasks, such as speech and character recognition, appropriately handling deformed patterns is a key to achieving high performance [Simard 92] [Simard 93] [Hinton 92] [Barnard 91]. The number of reasonable
deformations of patterns is enormous, since they can be either linear translations
(an affine translation or a time shifting) or non-linear deformations (a set of combinations of partial translations), or both.
Although a simple neural network (e.g. a 3-layered neural network) is able to adapt
Minoru Asogawa
1102
~'JU""'Y'CeUs
j-th Input Cell
/!/"....-I,ncul Field
.;_-;)oun~e
Image
Figure 1: AIFNN
ne~~----~----~~--------------
~----~--------------------~s
Position
Figure 2: Delta Force
non-linear deformations and to discriminate noises, it is still necessary to have
additional methods or data to appropriately process deformations.
This paper presents a new type of classification system, an Adaptive Input Field
Neural Network (AIFNN), which includes a simple pre-trained neural network and
an elastic input field attached to an input layer . The neural network is applied to
non-linear deformation compensations and the elastic input field to linear deformations .
The AIFNN algorithm can determine an optimal affine translation for compensating
for the original patterns' deformations, which are misclassified by the pre-trained
neural network. As the result, those misclassified patterns are correctly classified
and the final classification performance is improved, compared to that for the original neural network, without modifying the neural network.
2
Adaptive Input Field Neural Network (AIFNN)
AIFNN includes a pre-trained neural network and an elastic input field attached to
an input layer (Fig. 1) . The elastic input field contains receptors sampling input
patterns at each location . Each receptor connects to a cell in the input layer. Each
receptor links to its adjacent receptors with an elastic constraint and can move over
Adaptive Elastic Input Field for Recognition Improvement
1103
the input pattern independently, as long as its relative elastic constraint is satisfied.
The affine translation of the whole receptor (e.g. a shift, rotation, scale and slant
translation) satisfies an elastic constraint, since a constraint violation is induced by
the receptors' relative locations. 1 Partial deformations are also allowed with a
little constraint violation .
This feature of the elastic constraint is similar to that of the Elastic Net method
[Durbin 87], which can solve NP-hard problems. Although this elastic net method
is directly applicable to the template matching method, the performance is highly
dependent on the template selection. Therefore, an elaborated feature space for
non-linear deformations is mandatory [Hinton 92]. AIFNN utilizes something like
an elastic net constraint, but does not require any prominent templates.
The AIFNN algorithm is a repeated sequence of a bottom-up process (calculating
a guess and comparing with the presumption) and a top-down process (modifying
receptor's location to decrease the error and to satisfy the input field constraints).
For applying AIFNN as a classifier, a parallel search is performed; all categories
are chosen as presumption categories and the AIFNN algorithm is executed. After
hundreds of repetitions, an L score is calculated, which is the sum of the error and
the constraint violation in the elastic input field. A category which produces the
lowest L score is chosen as a plausible category. In Section 3, it is proved that all
receptors will settle to an equilibrium state. In the following sections, details about
the bottom-up and top-down processes are described.
Bottom-Up Process:
When a novel pattern is presented, each receptor samples activation corresponding
to a pattern intensity at each location. Each receptor activation is directly transmitted to a corresponding neural network input cell. Those input values are forwarded
through a pre-trained neural network and an output guess is obtained.
This guess is compared to the presumption category, and the negative of this error
is defined as the presumption certainty. 2 For example, using the mean squared
error criterion, the error ED is defined as follows;
ED
= ~ I)dk -
Ok?,
(1)
k
where Ok is the output value, and dk is the desired value determined by the presumption category. The presumption certainty is defined as _ED.
Top-Down Process:
To minimize the error and to maximize the presumption certainty, each receptor
modifies the activation by moving its location over the input pattern. The new
location for each receptor is determined by two elements; a direction which yields
less error and a direction which satisfies the input field elastic constraint. The
former element is called a Delta Force, since it relates to a delta value of an input
layer cell. The latter element is named an Address Force. Each receptor moves to
1 In previous papers, [Asogawa 90] and [Asogawa 91], a shift and rotation translation
was taken into account. In those models, a scale and slant translation violated the elastic
constraint.
2 Although another category coding schema is also possible, for simplicity, it is presumed
that each output cell corresponds to one certain category.
1104
Minoru Asogawa
a new location, which is determined by a sum of those two forces. The sum force is
called the Combined Force. In the next two sections, details about these forces are
described.
Delta Force: The Delta Force, which reduces ED by altering receptors' locations,
is determined by two elements: a partial derivative for the input value to the error,
and a local pattern gradient at each receptor location (Fig. 1).
To decrease the error ED, the value divergence for the j-th cell is computed as,
!lnet ?
}
D {)E D
D
== -a - = a 6? ,
onetj
}
(2)
where aD is small positive number and 6j is a delta value for the j-th input cell
and computed by the back-propagation [Yamada 91]. !lnetj and a local pattern
g~adient \1,pj are utilized to calculate a Delta Force !lsf; a scalar value of !lsf is
given as,
D
!lnetj
l!lsj I = 1\1,pj I .
Force !lsf is chosen as
The direction of the Delta
Consequently, !lsf is given as,
D _
!lnetj \1,pj _
being parallel to that of \1,pj .
6j
D
!lSj - 1\1,pjll\1,pjl -
(3)
1\1,pjI2 \1,pj .
a
(4)
To avoid !lsf becoming infinity, when 1\1,pi I is almost equal to 0, a small constant
c(
is added to the denominator; therefore, !lsf is defined as,
= t)
D _
!lSj -
a
D
6j
.
1\1,pj 12 + c\1,p} .
(5)
Address Force: If each receptor is moved iteratively following only the Delta Force,
the error becomes its minimum. However, receptors may not satisfy the input field
constraint and induce a large constraint violation EA. Here, EA is defined by a
distance between a receptor's lattice S and a lattice which is derived by an affine
translation from the original lattice. Therefore, EA is defined as follows;
E
1
A
N
2"d(S, S) =
2"1",
L.,.-IlsiN - sill 2
i
1
0
(6)
"2d(T(S ; t), S),
where d(?,?) is a distance measure for two receptor's lattices. S is a current receptor lattice . SN is the receptor lattice given by the affine translation T (.) with
parameters t and SO. SO is the original receptor lattice.
Therefore, as long as the receptor's lattice can be driven by some affine translation,
there is no constraint violation.
The affine parameters t are estimated so as to minimize EA;
{)EA
-{)- = 0
ti
for i
= 1, ? ??,6.
(7)
Adaptive Elastic Input Field for Recognition Improvement
1105
Since EA is quadratic with respect to ti, computing ti is moderate. The Address
Force for the j-th receptor ~st is defined as the partial derivative to EA with
respect to the receptor's location Sj;
A_
~Sj = -():
Aa;:-'
aEA
(8)
1
where
(}:A
is a small positive constant.
Combined Force: Here, all receptors are moved by a Combined Force
a sum of the Delta Force ~sD and the Address Force ~sA.
~s,
which is
After one hundred iterations, all receptors are moved to the location which produces
the minimum output error and the minimum constraint violation. Final states are
evaluated with a new measurement score, which is the sum of the error
and
the constraint violation
i.e.
+
L
L= ED EA.
EA;
ED
This L score is utilized to choose the correct category in a parallel search. In a parallel search, each category is temporarily chosen as a presumption and converged
L scores are calculated. Those scores are compared and the category yielding the
smallest L score is chosen as the correct category. This method fully exploits the
features of AIFNN, but it requires a large amount of computation, which can fortunately be processed totally in parallel. In the following section, convergence of the
AIFNN is shown.
3
Convergence
Convergence is shown by proving that the L is a Lyapunov function. When the L is
a Lyapunov function, all receptors converge to some locations after iterations. The
necessary and sufficient conditions for a Lyapunov function are (1) L has a lower
bound and (2) L monotonically decreases by applying the Combined Forces.
(12 Lower Bound:
E is the squared error at the output layer. Therefore, ED ~ O. EA is the constraint violation, which is defined with a distance between two lattices. Therefore,
EA ~ O. Since the L is a sum of ED and EA, the existence of a lower bound for
the L is proved.
0
(2) Monotonically Decrease:
The derivative of the L is calculated to show that the L decreases monotonically.
dL
d
d
dt
a:t+dT
ED
EA
I:. { aE
aS
,
D
i
A
~}
+ I:
{aE
dt
.
aS ~}
dt
,
i
~{(a::
+ aa~:) ~~i},
,
(9)
d Si IS
. t he Comb'me d Force an d gIven
.
h
were
dt
as,
d
s I? =
__
dt
~sD
+ ~SA.
(10)
Minoru Asogawa
lJ06
When a source image is smooth and 1'V?d is smaller than c, the following approximation is satisfied;
(11)
By using Eq. (11), the Delta Force is approximated as follows;
6.sD
=
aD
'V ?i
6i ~ _aD OED.
1\7?d 2 + c 'V ?i
OSi
(12)
By using Eqs. (8) and (12), and by letting aD = a A , the L derivative is computed
as follows;
dL
dt
......
OED
_a A ' " ( Li
OEA)2
as? + -as? ~ O.
I
(13)
I
With Eq. (13), it is proved that L decreases monotonically.D
4
Experiments and Results
Hand-written numerals recognition is chosen as one of the applications of AIFNN,
since performance improvement is shown by compensating for deformations
[Simard 92] [Simard 93] [Hinton 92] . The numeral inputs are bi-Ievel images of
32x40. They are blurred with a 5x5 Gaussian kernel and resampled to 14x 18 pixel
gray level images. To calculate an intensity and a local gradient between grids,
bi-linear Lagrange interpolation is utilized.
A neural network is 3 layered. The numbers of cells for the input layer, the hidden layer and the output layer are 252, 20 and 10, respectively. To obtain a
simpler weight configuration, two techniques are utilized; a constant weight decay [Ishikawa 89] and a small constant addition to output function derivatives
[Fahlman 88]. Training is repeated for 180 epochs with 2500 numerals, and tested
with another 2500. Since image edges are almost blank, about 2400 connections
between the input layer and the hidden layer are equal to 0; therefore, the number
of parameters is reduced to 2870.
In this experiment, a simple decision method is used; the maximum output cell is
chosen as a guess and patterns are rejected when the error of the guess is greater
than a threshold value . Naturally, a low threshold yields a low misclassification
rate, but also yields a high rejection rate [Martin 92]. With the maximum threshold, the rates of rejection, correct classification and misclassification are 0.00%(0
patterns), 95.20%(2380 patterns) and 4.80%(120 patterns), respectively. For the
2500 numerals learning data, these rates are 0.00%(0 patterns), 99.40%(2485 patterns) and 0.60%(15 patterns). When a threshold is 0.001, the rates of rejection,
correct classification and misclassification are 43.52%(1088 patterns), 56.40%(1410
patterns) and 0.08%(2 patterns), respectively.
AIFNN is applied to these 1088 rejected patterns. and classifies 997 patterns correctly. Therefore, total performances for rejection, correct classification and misclassification become 0.00%(0 patterns), 95.72%(2393 patterns) and 4.28%(107 patterns), respectively. As the classification performance is improved; the number of
Adaptive Elastic Input Field for Recognition Improvement
1107
misclassified patterns reduces from 120 to 107 without modifying the neural network. 10.83% of the originally misclassified patterns are correctly categorized. Fig.
3 shows an input field after one hundred iterations.
.....
.......
......
......
??????
.
????? ?
6
?
.. "
?
III!
?
0.111
?
It . . . .
?atOl" ??
??????
?????
???
......
'??????
.........?iiI". .....?.
?
A ....
)0
,.
,
I::
I
18
?
???
6
it
.
"
. .y............
Ilttput Activation
I... ....... . ... 1. .
0123456789
In the figure on the left, receptors are located at each grid point in a gray lattice. The
circle diameter corresponds to the pattern intensity at each receptor's location. The bottom right figure indicates the source image, and the top right figure indicates the neural
network input. This image was initially misclassified as 3 instead of 8. After iteration
with presumption as 8, category 8 gets the highest activation and the receptor's lattice is
rotated to compensate for the initial deformation.
Figure 3: Input Field After Adaptation
5
Discussion
It is shown that the AIFNN can improve the classification performance for the
original neural network, without modifications. This performance improvement is
caused by an optimal affine translations estimation for rejected patterns.
Although an affine translation is discussed in this paper, the algorithm is applicable to any deformation mechanism; such as a gain and offset equalization and 3D
perspective deformation.
1108
Minoru Asogawa
The requirement for a neural network in AIFNN is the capability of calculating
partial derivatives for an input layer, so a layered neural network is utilized in
this paper. Since partial derivative can be computed by numerical approximation,
practically any neural network is applicable for AIFNN. Moreover, any differentiable
error criterion is applicable; such as, a KL information and a likelihood.
To reduce computation, a sequential searching is also possible; a presumption is
chosen as the most plausible category, e.g. the smallest error category. If the L
score falls behind a threshold, this presumption is regarded as correct. If it's not,
another plausible category is chosen as a presumption and tested [Asogawa 91] .
References
[As ogawa 90] M. Asogawa, "Adaptive Input Field Neural Network - that can recognize rotated and/or shifted character -", Proceedings of IJCNN '90 at San
Diego, vol. 3. pp. 733-738. June 1990.
[As ogawa 91] M. Asogawa, "Adaptive Input Field Neural Network", Proceedings of
IJCNN '91 at Singapore, vol. 1. pp. 83-88. November 1991.
[Barnard 91] E. Barnard et aI., "Invariance and Neural Nets" , IEEE trans. on Neural Networks, vol. 2. no. 5, pp . 498-508. 1992.
[Durbin 87] R. Durbin et al., "An analogue approach to the traveling salesman
problem using an elastic net method", Nature, vol. 326. pp. 689-691. 1987.
[Fahlman 88] S. Fahlman, "An empirical study of learning speed in backpropagation networks", CMU-CS-88-162, 1988.
[Hinton 92] G .E. Hinton et al., "Adaptive Elastic Models for Hand-Printed Character Recognition", Advances in Neural Information Processing Systems, vol.
4. pp. 512-519. 1992.
[Ishikawa 89] M. Ishikawa, "A structural learning algorithm with forgetting of link
weights" , Proceedings of IJCNN '89 at Washington DC., vol. 2, pp. 626, 1989.
[Martin 92] G. L. Martin et al., "Recognizing Overlapping Hand-Printed Characters by Centered-Object Integrated Segmentation and Recognition", Advances
in Neural Information Processing Systems, vol. 4. pp. 504-511. 1992.
[Simard 92] P. Simard et al., "Tangent Prop - A Formalism for Specifying Selected
Invariances in an Adaptive Network", Advances in Neural Information Processing Systems, vol. 4. pp. 895-903. 1992.
[Simard 93] P . Simard et al., "Efficient Pattern Recognition Using a New Transformation Distance" , Advances in Neural Information Processing Systems, vol. 5.
pp. 50-58. 1993.
[Yamada 91] K . Yamada, "Learning of category boundaries based on inverse recall
by multilayer neural network", Proceedings of IJCNN '91 at Seattle, pp. 7-12
vol.2 1991.
| 969 |@word deformed:2 initial:1 configuration:1 contains:1 score:8 current:1 comparing:1 blank:1 activation:5 si:1 written:1 numerical:1 selected:1 guess:5 yamada:3 location:13 simpler:1 become:1 comb:1 forgetting:1 presumed:1 compensating:2 csl:1 little:1 totally:1 becomes:1 classifies:1 moreover:1 lowest:1 miyazaki:1 transformation:1 certainty:3 ti:3 classifier:1 positive:2 local:3 sd:3 receptor:31 becoming:1 interpolation:1 specifying:1 co:1 sill:1 bi:2 backpropagation:1 empirical:1 printed:2 matching:1 pre:5 induce:1 get:1 layered:3 selection:1 applying:2 equalization:1 adient:1 modifies:1 independently:1 simplicity:1 regarded:1 proving:1 searching:1 diego:1 element:4 recognition:11 approximated:1 utilized:5 located:1 bottom:4 calculate:2 decrease:6 highest:1 oed:2 trained:5 solve:1 plausible:3 forwarded:1 final:2 sequence:1 differentiable:1 net:5 adaptation:1 osi:1 moved:3 ogawa:2 seattle:1 convergence:4 requirement:1 produce:2 rotated:2 object:1 eq:3 sa:2 c:1 lyapunov:3 direction:3 correct:6 modifying:4 centered:1 settle:1 numeral:5 require:1 practically:1 equilibrium:1 smallest:2 estimation:1 applicable:4 repetition:1 gaussian:1 avoid:1 derived:1 june:1 improvement:6 indicates:2 likelihood:1 dependent:1 integrated:1 initially:1 hidden:2 misclassified:6 pixel:1 classification:10 field:22 equal:2 washington:1 sampling:1 ishikawa:3 np:1 divergence:1 recognize:1 connects:1 highly:1 violation:8 yielding:1 behind:1 edge:1 partial:6 necessary:2 desired:1 circle:1 deformation:14 formalism:1 altering:1 lattice:11 hundred:3 recognizing:1 accomplish:1 combined:4 ju:1 st:1 squared:2 satisfied:2 choose:1 simard:8 derivative:7 li:1 japan:1 account:1 coding:1 includes:3 blurred:1 satisfy:2 caused:1 ad:4 performed:1 schema:1 parallel:5 capability:1 elaborated:1 minimize:2 yield:3 handwritten:1 classified:1 converged:1 ed:10 pp:10 naturally:1 gain:1 proved:3 recall:1 segmentation:1 ea:13 back:1 ok:2 originally:2 dt:7 improved:3 evaluated:1 rejected:3 traveling:1 hand:3 overlapping:1 propagation:1 gray:2 former:1 laboratory:1 iteratively:1 adjacent:1 x5:1 criterion:2 prominent:1 image:7 novel:1 rotation:2 attached:3 jp:1 discussed:1 he:1 measurement:1 slant:2 ai:1 grid:2 moving:1 something:1 perspective:1 moderate:1 driven:1 mandatory:1 certain:1 transmitted:1 minimum:3 additional:1 fortunately:1 greater:1 converge:1 determine:2 maximize:1 monotonically:4 relates:1 reduces:2 smooth:1 adapt:1 compensate:2 long:2 denominator:1 ae:2 cmu:1 multilayer:1 iteration:4 kernel:1 cell:9 addition:1 lsf:6 source:2 appropriately:3 induced:1 structural:1 iii:2 reduce:1 shift:2 x40:1 aea:1 speech:2 ievel:1 amount:1 processed:1 category:17 diameter:1 reduced:1 singapore:1 shifted:1 delta:11 estimated:1 correctly:4 vol:10 key:2 threshold:5 enormous:1 achieving:2 pj:6 sum:6 inverse:1 named:1 almost:2 reasonable:1 utilizes:1 decision:1 layer:12 bound:3 resampled:1 quadratic:1 durbin:3 ijcnn:4 constraint:17 infinity:1 speed:1 martin:3 combination:1 smaller:1 character:5 modification:1 taken:1 mechanism:1 letting:1 salesman:1 existence:1 original:6 top:4 calculating:2 exploit:1 move:2 added:1 minoru:5 gradient:2 distance:4 link:2 me:1 executed:1 negative:1 perform:1 compensation:1 november:1 hinton:5 dc:1 intensity:3 kl:1 connection:1 nee:1 trans:1 address:4 able:1 pattern:33 shifting:1 analogue:1 misclassification:4 force:21 improve:1 ne:1 sn:1 epoch:1 tangent:1 relative:2 fully:1 presumption:12 affine:10 sufficient:1 pi:1 translation:14 fahlman:3 fall:1 template:3 boundary:1 calculated:3 author:1 adaptive:11 san:1 lsj:3 sj:2 a_:1 search:3 iterative:1 nature:1 pjl:1 kanagawa:1 elastic:22 cl:1 whole:1 noise:1 allowed:1 repeated:2 categorized:2 fig:3 miyamae:1 position:1 oun:1 down:3 offset:1 dk:2 decay:1 dl:2 sequential:1 nec:1 rejection:4 lagrange:1 temporarily:1 scalar:1 aa:2 corresponds:2 satisfies:2 prop:1 consequently:2 barnard:3 hard:1 determined:4 total:2 called:2 discriminate:1 invariance:2 latter:1 violated:1 kawasaki:1 tested:2 handling:2 |
7,209 | 97 | 289
NEURAL ANALOG DIFFUSION-ENHANCEMENT
LAYER AND SPATIO-TEMPORAL GROUPING
IN EARLY VISION
Allen M. Waxman?,t, Michael Seibert?,t,RobertCunninghamt and I ian Wu?
? Laboratory for Sensory Robotics
Boston University
Boston, MA 02215
t Machine Intelligence Group
MIT Lincoln Laboratory
Lexington, MA 02173
ABSTRACT
A new class of neural network aimed at early visual processing is
described; we call it a Neural Analog Diffusion-Enhancement Layer or
"NADEL." The network consists of two levels which are coupled
through feedfoward and shunted feedback connections. The lower level
is a two-dimensional diffusion map which accepts visual features as
input, and spreads activity over larger scales as a function of time. The
upper layer is periodically fed the activity from the diffusion layer and
locates local maxima in it (an extreme form of contrast enhancement)
using a network of local comparators. These local maxima are fed back
to the diffusion layer using an on-center/off-surround shunting
anatomy. The maxima are also available as output of the network. The
network dynamics serves to cluster features on multiple scales as a
function of time, and can be used in a variety of early visual processing
tasks such as: extraction of comers and high curvature points along
edge contours, line end detection, gap filling in contours, generation of
fixation points, perceptual grouping on multiple scales, correspondence
and path impletion in long-range apparent motion, and building 2-D
shape representations that are invariant to location, orientation, scale,
and small deformation on the visual field.
INTRODUCTION
Computer vision is often divided into two main stages, "early vision" and "late vision",
which correspond to image processing and knowledge-based recognition/interpretation,
respectively. Image processing for early vision involves algorithms for feature
enhancement and extraction (e.g. edges and comers), feature grouping (i.e., perceptual
We acknowledge support from the Machine Intelligence Group of MIT Lincoln
Laboratory. The views expressed are those of the authors and do not reflect the official
policy or position of the U.S. Government
290
Waxman, Seibert, Cunningham and Wu
organization), and the extraction of physical properties for object surfaces that comprise a
scene (e.g. reflectance, depth, surface slopes and curvatures, discontinuities). The
computer vision literature is characterized by a plethora of algorithms to achieve many of
these computations, though they are hardly robust in performance.
Biological neural network processing does, of course, achieve all of these early vision
tasks, as evidenced by psychological studies of the preattentive phase of human visual
processing. Often, such studies provide motivation for new algorithms in computer
vision. In contrast to this algorithmic approach, computational neural network processing
tries to glean organizational and functional insights from the biological realizations, in
order to emulate their information processing capabilities. This is desirable. mainly
because of the adaptive and real-time nature of the neural network architecture. Here, we
shall demonstrate that a single neural architecture based on dynamical diffusionenhancement networks can realize a large variety of early vision tasks that deal mainly
with perceptual grouping. The ability to group image features on multiple scales as a
function of time, follows from "attractive forces" that emerge from the network
dynamics. We have already implemented the NADEL (in 16-bit arithmetic) on a videorate parallel computer, the PIPE [Kent et al .? 1985], as well as a SUN-3 workstation.
THE NADEL
The Neural Analog Diffusion-Enhancement Layer was recently introduced by Seibert &
Waxman [1989]. and is illustrated in Figure 1; it consists primarily of two levels which
are coupled via feedforward and shunted feedback connections. Low-level features
extracted from the imagery provide input to the lower level (a 2-D map) which spreads
input activity over larger scales as time progresses via diffusion, allowing for passive
decay of activity. The diffused activity is periodically sampled and passed upward to a
contrast-enhancing level (another 2-D map) which locates local maxima in the terrain of
diffuse activity. However, this forward pathway is masked by receptive fields which pass
only regions of activity with positive Gaussian curvature and negative mean curvature;
that is these receptive fields play the role of inhibitory dendro-dendritic modulatory gates.
This masking fascilitates the local maxima detection in the upper level. These local
maxima detected by the upper level are fed back to the lower diffusion level using a
shunting dynamics with on-center/off-surround anatomy (cf. [Grossberg, 1973] on the
importance of shunting for automatic gain control, and the role of center/surround
anatomies in competitive networks). The local maxima are also available as outputs of the
network, and take on different interpretations as a function of the input. A number of
examples of spatio-temporal grouping will be illustrated in the next section.
The primary result of diffusion-enhancement network dynamics is to create a longrange attractive force between isolated featural inputs. This force manifests itself by
shifting the local maxima of activity toward one another. leading to a featural grouping
over mUltiple scales as a function of time. This is shown in Figure 1, where two featural
inputs spread their initial excitations over time. The individual activities superpose. with
the tail of one gaussian crossing the maximum of the other gaussian at an angle. This
Neural Analog Diffusion-Enhancement Layer
biases the superposition of activities, adding more activity to one side of a maximum than
another, causing a shift in the local maxima toward one another. Eventually, the local
maxima merge into a single maximum at the centroid of the individual inputs. If we keep
track of the local maxima as diffusion progresses (by connecting the output of the
enhancement layer to another layer which stores activity in short term memory), then the
two initial inputs will become connected by a line. In Figure 1 we also illustrate the
grouping of five features in two clusters, a configuration possessing two spatial scales.
After little diffusion the local maxima are located where the initial inputs were. Further
diffusion causes each cluster to form a single local maximum at the cluster centroid.
Eventually, both clusters merge into a single hump of activity with one maximum at the
centroid of the five initial inputs. Thus, multi scale grouping over time emerges. The
examples of Figure 1 use only diffusion without any feedback, yet they illustrate the
importance of localizing the local maxima through a kind of contrast-enhancement on
another layer. The local maxima of activity serve as "place tokens" representing grouped
features at a particular scale. The feedback pathway re-activates the diffusion layer,
thereby allowing the grouping process to proceed to still larger scales, even across
featureless areas of imagery.
The dynamical evolution of activity in the NADEL can be modeled using a modified
diffusion equation [Seibert & Waxman, 1989]. However, in our simulations of the
NADEL we don't actually solve this differential equation directly. Instead, each iteration
of the NADEL consists of a spreading of activity using gaussian convolution, allowing
for passive decay, then sampling the diffusion layer, masking out areas which are not
positive Gaussian curvature and negative mean curvature activity surfaces, detecting one
local maximum in each of these convex areas, and feeding this back to the diffusion layer
with a shunted on-center/off-surround excitation at the local maxima. In the biological
system, diffusion can be accomplished via a recurrent network of cells with offcenter/on-surround lateral connectivity, or more directly using electrotonic coupling
across gap junctions as in the horizontal cell layer of the retina [Dowling, 1987].
Curvature masking of the activity surface can be accomplished using oriented offcenter/on-surround receptive fields that modulate the connections between the two
primary layers of the NADEL.
SPATIO-TEMPORAL GROUPING
We give several examples of grouping phenomena in early vision, utilizing the NADEL.
In all cases its parameters correspond to gaussian spreading with 0'=3 and passive decay
of 1% per iteration, and on-center/off-surround feedback with 0'+=11'12 and 0'_=1.
Grouping of Two Points: The simple case of two instantaneous point stimuli input
simultaneously to the NADEL is summarized in Figure 2. We plot the time (N network
iterations) it takes to merge the two inputs, as a function of their initial separation (S
pixels). For S:S;6 the points merge in one iteration; for S>24 activity equilibrates and
shifting of local maxima never begins.
291
292
Waxman, Seibert, Cunningham and Wu
Grouping on Multiple Scales: Figure 3 illusttates the hierarchy of groupings generated
by a square outline (31 pixels on a side) with gaps (9 pixels wide). Comer and line-end
features are first enhanced using complementary center-surround receptive fields
(modeled as a rectified response to a difference-of-gaussians), and located at the local
maxima of activity. These features are shown superimposed on the shape in 3a; they
serve as input to the NADEL. Figure 3b shows the loci of local maxima determined up to
the second stable grouping, superimposed over the shape. Boundary completion fills the
gaps in the square. In Figure 3c we show the loci of local maxima on the image plane,
after the final grouping has occured (N=I00 iterations). The trajectory of local maxima
through space-time (x,y,t) is shown in Figure 3d after the fourth grouping. It reveals a
hierarchical organization similar to the "scale-space diagrams" of Witkin [1983].
It can be seen from Figure 3d that successive groupings form stable entities in that the
place tokens remain stationary for several iterations of the NADEL. It isn't until activity
has diffused farther out to the next representative scale that these local maxima start
moving once again, and eventually merge. This relates stable perceptual groupings to
place tokens (i.e., local maxima of activity) that are not in motion on the diffusion layer.
The motion of place tokens can be measured in the same fashion as feature point motion
across the visual field. Real-time receptive fields for measuring the motion of image edge
and point features have recently been developed by Waxman et ale [1988].
Grouping of Time-Varying Inputs: The simplest example in this case corresponds to the
grouping of two lights that are flashed at different locations at different times. When the
time interval between flashes (Stimulus Onset Asynchrony SOA) is set appropriately,
one perceives a smooth motion _or "path impletion" between the stimuli. This percept of
"long-range apparent motion" is the cornerstone of the Gestalt Psychology moverment,
and has remained unexplained for one-hundred years now [Kolers, 1972]. We have
applied the NADEL to a variety of classical problems in apparent motion including the
"split motion" percept and the mUlti-point Ternus configuration [Waxman et al., 1989].
Here we consider only the case of motion between two stimuli, where we interpret the
locus of local maxima as the impleted path in apparent motion. However, the direction
of perceived motion is not determined by the grouping process itself; only the path. We
make the additional assumption that grouping generates a motion percept only if the
second stimulus begins to shift immediately upon input to the NADEL. We suggest that
the motion percept occurs only after path impletion is complete. That is, while grouping
is active, its outputs are suppressed from our perception (a form of "ttansient-onsustained inhibition" analogous to saccadic suppression). By varying the separation
between the two stimuli, and the time (SOA) between their inputs, we can plot regimes
for which the NADEL predicts apparent motion. This is shown in Figure 4, which
compares favorably with the psychophysical results summarized in Figure 3.2 of [Kolers,
1972]. We find regimes in which complete paths are formed ("smooth motion"), partial
paths are formed ("jumpy motion"), and no immediate shifting occurs ("no motion"). The
maximum allowable SOA between stimuli (upper curves) is determined by the passive
decay rate. Increasing this decay from 1% to 3% will decrease the maximum SOA by a
Neural Analog Diffusion-Enhancement Layer
factor of five. The minimum allowable SOA (lower curves) increases with increasing
separation, since it takes longer for activity from the first stimulus to influence a more
distant second stimulus. The linearity of the lower boundary has been interpreted by
[Waxman et aI.? 1989] as suggestive of Korte's "third law" [Kolers, 1972], when taken in
combination with a logarithmic transformation of the visual field [Schwartz, 1980].
Attentional Cues and Invariant Representations: Place tokens which emerge as stable
groupings over time can also provide attentional cues to a vision system. They would
typically drive saccadic eye motions during scene inspection, with the relative activities
of these maxima and their order of emergence determining the sequence of rapid eye
motions. Such eye motions are known to play a key role in human visual perception
[Yarbus, 1967]; they are influenced by both bottom-up perceptual cues as well as topdown expectations. The neuromorphic vision system developed by Seibert & Waxman
[1989], shown in Figure 5, utilizes the NADEL to drive "eye motions", and thereby
achieve translational invariance in 2-D object learning and recognition. This is followed
by a log-polar transform (which emulates the geniculo-cortical connections [Schwartz,
1980]) and another NADEL to achieve rotation and scale in variance as well. Further
coding of the transformed feature points by overlapping receptive fields provides
invariance to small deformation. Pattern learning and recognition is then achieved using
an Adaptive Resonance Theory (ART-2) network [Carpenter & Grossberg, 1987].
REFERENCES
G. Carpenter & S. Grossberg (1987). ART-2: Self-organization of stable category
recognition codes for analog input patterns. Applied Optics 26, pp. 4919-4930.
I.E. Dowling (1987). The RETINA: An approachable part or the brain. Cambridge,
MA: Harvard University Press.
S. Grossberg (1973). Contour enhancement, short term memory, and constancies in
reverberating neural networks. Studies in Applied Mathematics 52. pp.217-257.
E.W. Kent, M.O. Shneier & R. Lumia (1985). PIPE: Pipelined Image Processing Engine.
Journal of Parallel and Distributed Computing 2, pp. 50-78.
P.A. Kolers (1972). Aspects or Motion Perception. New York: Pergamon Press.
E.L. Schwartz (1980). Computational anatomy and functional architecture of striate
cortex: A spatial mapping approach to perceptual coding. Vision Research 20, pp. 645669.
M. Seibert & A.M. Waxman (1989). Spreading activation layers, visual saccades and
invariant representations for neural pattern recognition systems. N~ural Networks 2, pp.
9-27.
293
294
Waxman, Seibert, Cunningham and Wu
A.M. Waxman, J. Wu & F. Bergholm (1988). Convected activation profiles and the
measurement of visual motion. Proceeds. 1988 IEEE Conference on Computer Vision
and Pattern Recognition. Ann Arbor, MI, pp. 717-723.
A.M. Waxman. J. Wu & M. Seibert (1989). Computing visual motion in the short and
the long: From receptive fields to neural networks. Proceeds. IEEE 1989 Workshop on
Visual Motion. Irvine, CA.
A.P. Witkin (1983). Scale space filtering. Proceeds. of the International Joint
Conference on Artificial Intelligence. Karlsruhe, pp. 1019-1021.
AL. Yarbus (1967). Eye Movements and Vision. New York: Plenum Press.
ON?CENTER!
o
o
o
OFF?SURROUND
OFF?CENTER!
a.l-SURAOUNQ
SHUNTNG
FEE08AO<
ON-CENTER!
OFF?SURAOUND
Figure 1 ? (left) The NADEL takes featural input and diffuses it over a 2-D map as a
function of time. Local maxima of activity are detected by the upper layer and fed back to
the diffusion layer using an on-center/off-surround shunting anatomy. (right. top)
Features spread their activities which superpose to generate an attractive force, causing
distant features to group. (right, bottom) Grouping progresses in time over multiple
scales. with small clusters emerging before extended clusters. Clusters are represented by
their local ~axima of activity, which serve as place tokens.
1ii~Ill'J!I!IIIl!lll!l1llllllm=
!,~
~l"
lmilllillUml:l ?
1_?
1000
600
.00
I;
200
/,
~
I
so
~
~~
..
I~~
Ii'
~,I
~
tt~i'><'!
k;'~
"~~~"'r..;.'(!S.1-:'~-'
~,~o.~_,
?
~7'~~':;'
~~~~ .
...
(b)
(a)
I
~o
/
.x
~
?
':'.~
~
~
/
60
.9
~
1
~h
t~5~
~~!:Ji.
.J\llt!
:??'!"Id
I
I
lOa
<!
1~~'l
F~
.'
;h~ '
800
z
~
e.
~
e.
/
/
20
I
/
10
I
I
6
I
t2
/
~
y
/
~
.---
.... ,-_ .....
r!3.
g
Separalion
~
5 (pIXelS)
g.
(e)
Cd)
I
g
Figure 2 _ The time (network iterations N) to merge
two points input simultaneously to the NADEL, as a
function of their initial separation (S pixels).
("t-
~
Figure 3 - Perceptual grouping of a square outline with gaps.
~
~
~
'-0
c:.n
~
co
(J')
E_pectat Ions
Spatial relations
200
80
g
,/
40
Smooth Motion
<II
::s
e.::s
./
./'
c:
.2
~
a>
,;+
No Motion
60
~
./
I?.
./
20
a
./'
-l:
,/
0
~
iii
.s
10
8
C
C/)
6
Jumpy Motion
~~
./ , /
./
~
./
./
./
4
/
/
/
2
/
VIA NADEL
/
1
J
....g-~
100
FEATURE MAP
6
Separation S (pixels)
Figure 4 - Apparent motion between two flashed
lights: Stimulus Onset Asynchrony SOA (network
iterations N) vs. Separation S (pixels). Solid curves
indicate boundaries between which introduction of
the second light yields immediate shifting of local
maxima; dashed curves (above solid curves) indicate
when final merge occurs yielding the impleted path.
EYE MOTIONS
Figure 5 - The neuromorphic vision system for
invariant learning and recognition of 2-D object~
utilizes three NADEL networks.
| 97 |@word simulation:1 kent:2 thereby:2 solid:2 initial:6 configuration:2 activation:2 yet:1 realize:1 distant:2 periodically:2 shape:3 plot:2 v:1 stationary:1 intelligence:3 cue:3 plane:1 inspection:1 feedfoward:1 short:3 farther:1 detecting:1 provides:1 location:2 successive:1 yarbus:2 five:3 along:1 become:1 differential:1 consists:3 fixation:1 pathway:2 rapid:1 multi:2 brain:1 little:1 lll:1 increasing:2 perceives:1 begin:2 linearity:1 kind:1 interpreted:1 emerging:1 developed:2 lexington:1 transformation:1 temporal:3 schwartz:3 control:1 positive:2 before:1 local:28 id:1 path:8 merge:7 jumpy:2 co:1 range:2 grossberg:4 area:3 suggest:1 pipelined:1 influence:1 map:5 center:10 convex:1 immediately:1 insight:1 utilizing:1 fill:1 analogous:1 plenum:1 enhanced:1 hierarchy:1 play:2 superpose:2 harvard:1 crossing:1 recognition:7 located:2 predicts:1 bottom:2 role:3 constancy:1 region:1 connected:1 sun:1 decrease:1 movement:1 dynamic:4 serve:3 upon:1 comer:3 joint:1 emulate:1 represented:1 detected:2 artificial:1 apparent:6 larger:3 solve:1 ability:1 emergence:1 itself:2 transform:1 final:2 sequence:1 causing:2 realization:1 achieve:4 lincoln:2 enhancement:11 cluster:8 plethora:1 object:3 illustrate:2 recurrent:1 coupling:1 completion:1 measured:1 progress:3 implemented:1 involves:1 indicate:2 direction:1 anatomy:5 human:2 government:1 feeding:1 biological:3 dendritic:1 algorithmic:1 mapping:1 early:8 perceived:1 polar:1 spreading:3 superposition:1 unexplained:1 grouped:1 create:1 mit:2 activates:1 gaussian:6 modified:1 varying:2 superimposed:2 mainly:2 contrast:4 centroid:3 suppression:1 typically:1 cunningham:3 diffuses:1 relation:1 transformed:1 upward:1 pixel:7 translational:1 orientation:1 ill:1 resonance:1 spatial:3 art:2 iiil:1 field:10 comprise:1 never:1 extraction:3 once:1 sampling:1 comparators:1 filling:1 t2:1 stimulus:10 primarily:1 retina:2 oriented:1 simultaneously:2 approachable:1 individual:2 phase:1 detection:2 organization:3 hump:1 extreme:1 yielding:1 light:3 edge:3 partial:1 re:1 deformation:2 isolated:1 psychological:1 localizing:1 measuring:1 neuromorphic:2 organizational:1 masked:1 hundred:1 international:1 off:8 michael:1 shunted:3 connecting:1 connectivity:1 imagery:2 reflect:1 again:1 equilibrates:1 leading:1 waxman:13 summarized:2 coding:2 onset:2 view:1 try:1 competitive:1 start:1 capability:1 parallel:2 masking:3 slope:1 square:3 formed:2 variance:1 emulates:1 percept:4 correspond:2 yield:1 trajectory:1 rectified:1 drive:2 llt:1 influenced:1 pp:7 mi:1 workstation:1 sampled:1 gain:1 irvine:1 manifest:1 knowledge:1 emerges:1 occured:1 actually:1 back:4 response:1 though:1 stage:1 until:1 horizontal:1 overlapping:1 asynchrony:2 karlsruhe:1 building:1 evolution:1 laboratory:3 flashed:2 illustrated:2 deal:1 attractive:3 during:1 self:1 excitation:2 allowable:2 outline:2 complete:2 demonstrate:1 tt:1 allen:1 motion:31 passive:4 image:6 instantaneous:1 recently:2 possessing:1 rotation:1 functional:2 physical:1 ji:1 analog:6 interpretation:2 tail:1 interpret:1 measurement:1 dowling:2 surround:10 cambridge:1 ai:1 automatic:1 mathematics:1 moving:1 stable:5 longer:1 surface:4 inhibition:1 cortex:1 curvature:7 store:1 accomplished:2 seen:1 minimum:1 additional:1 dashed:1 arithmetic:1 relates:1 ale:1 desirable:1 ii:3 witkin:2 multiple:6 ural:1 smooth:3 characterized:1 long:3 divided:1 shunting:4 locates:2 nadel:20 vision:16 enhancing:1 expectation:1 iteration:8 robotics:1 cell:2 ion:1 achieved:1 interval:1 diagram:1 appropriately:1 call:1 feedforward:1 split:1 iii:1 variety:3 geniculo:1 psychology:1 architecture:3 shift:2 passed:1 proceed:1 cause:1 hardly:1 york:2 electrotonic:1 cornerstone:1 modulatory:1 korte:1 aimed:1 category:1 simplest:1 generate:1 inhibitory:1 track:1 glean:1 per:1 shall:1 group:4 key:1 diffusion:22 year:1 angle:1 fourth:1 place:6 wu:6 separation:6 utilizes:2 bit:1 layer:20 followed:1 correspondence:1 activity:27 optic:1 scene:2 diffuse:1 generates:1 aspect:1 combination:1 across:3 remain:1 suppressed:1 invariant:4 taken:1 equation:2 eventually:3 locus:3 fed:4 serf:1 end:2 soa:6 available:2 junction:1 gaussians:1 hierarchical:1 gate:1 top:1 cf:1 reflectance:1 classical:1 psychophysical:1 diffused:2 pergamon:1 already:1 occurs:3 receptive:7 primary:2 saccadic:2 striate:1 attentional:2 lateral:1 entity:1 toward:2 code:1 modeled:2 favorably:1 negative:2 policy:1 allowing:3 upper:5 convolution:1 acknowledge:1 immediate:2 extended:1 introduced:1 evidenced:1 pipe:2 connection:4 offcenter:2 engine:1 accepts:1 discontinuity:1 topdown:1 dynamical:2 perception:3 i00:1 pattern:4 proceeds:3 regime:2 including:1 memory:2 shifting:4 force:4 representing:1 eye:6 coupled:2 featural:4 isn:1 literature:1 determining:1 relative:1 law:1 generation:1 filtering:1 cd:1 course:1 token:6 loa:1 bias:1 side:2 wide:1 emerge:2 distributed:1 feedback:5 depth:1 boundary:3 curve:5 cortical:1 contour:3 sensory:1 author:1 forward:1 adaptive:2 gestalt:1 longrange:1 keep:1 suggestive:1 active:1 reveals:1 spatio:3 terrain:1 don:1 nature:1 robust:1 ca:1 official:1 spread:4 main:1 motivation:1 featureless:1 profile:1 complementary:1 carpenter:2 representative:1 fashion:1 position:1 perceptual:7 late:1 third:1 ian:1 remained:1 reverberating:1 decay:5 grouping:27 workshop:1 adding:1 importance:2 gap:5 boston:2 logarithmic:1 visual:12 expressed:1 saccade:1 corresponds:1 extracted:1 ma:3 modulate:1 ann:1 flash:1 seibert:9 determined:3 pas:1 invariance:2 arbor:1 preattentive:1 support:1 phenomenon:1 |
7,210 | 970 | Sample Size Requirements For
Feedforward Neural Networks
Michael J. Turmon
Cornell Univ. Electrical Engineering
Ithaca, NY 14853
mjt@ee.comell.edu
Terrence L. Fine
Cornell Univ. Electrical Engineering
Ithaca, NY 14853
tlfine@ee.comell.edu
Abstract
We estimate the number of training samples required to ensure that
the performance of a neural network on its training data matches
that obtained when fresh data is applied to the network. Existing
estimates are higher by orders of magnitude than practice indicates.
This work seeks to narrow the gap between theory and practice by
transforming the problem into determining the distribution of the
supremum of a random field in the space of weight vectors, which
in turn is attacked by application of a recent technique called the
Poisson clumping heuristic.
1
INTRODUCTION AND KNOWN RESULTS
We investigate the tradeofi"s among network complexity, training set size, and statistical performance of feedforward neural networks so as to allow a reasoned choice
of network architecture in the face of limited training data. Nets are functions
7](x; w), parameterized by their weight vector w E W ~ Rd , which take as input
points x E Rk. For classifiers, network output is restricted to {a, 1} while for forecasting it may be any real number. The architecture of all nets under consideration
is N, whose complexity may be gauged by its Vapnik-Chervonenkis (VC) dimension
v, the size of the largest set of inputs the architecture can classify in any desired way
('shatter'). Nets 7] EN are chosen on the basis of a training set T
{(Xi, YiHr=l.
These n samples are i.i.d. according to an unknown probability law P. Performance
of a network is measured by the mean-squared error
E(w)
E(7](x; w) - y)2
(1)
= P(7](x;w);/; y) (for classifiers)
(2)
=
328
Michael Turman, Terrence L. Fine
and a good (perhaps not unique) net in the architecture is WO = argmiIlwew ?(w).
To select a net using the training set we employ the empirical error
1 n
VT(W) = - I)11(Xi; w) - Yi)2
(3)
n i=l
sustained by 11(?; w) on the training set T. A good choice for a classifier is then
w? = argmiIlwew VT(W). In these terms, the issue raised in the first sentence ofthe
section can be restated as, "How large must n be in order to ensure ?(w?)-?(WO) $
i with high probability?"
For purposes of analysis we can avoid dealing directly with the stochastically chosen
network w? by noting
?(w?) - ?(WO) $ IVT(W?) - ?(w?)1
+ IVT(WO) -
?(wo)1 $ 2 sup IVT(W) - ?(w)1
wEW
A bound on the last quantity is also useful in its own right.
The best-known result is in (Vapnik, 1982), introduced to the neural network community by (Baum & Haussler, 1989):
(2n)V
~
P( sup IVT(W) - ?(w)1 ~ i) $ 6-,-e- n ( /2
(4)
wEW
v.
This remarkable bound not only involves no unknown constant factors, but holds
independent of the data distribution P . Analysis shows that sample sizes of about
nc
= (4V/i 2) log 3/i
(5)
are enough to force the bound below unity, after which it drops exponentially to
zero. Taking i = .1, v = 50 yields nc = 68000, which disagrees by orders of
magnitude with the experience of practitioners who train such simple networks.
More recently, Talagrand (1994) has obtained the bound
K2ni2)v
~
P( sup IVT(W) - ?(w)1 ~ i) $ Kl (
e- 2n ( ,
(6)
wew
V
yielding a sufficient condition of order V/i 2, but the values of Kl and K2 are inaccessible so the result is of no practical use.
Formulations with finer resolution near ?(w) = 0 are used. Vapnik (1982) bounds
P(suPwew IVT(W) - ?(w)I/?(w)1/2 ~ i)-note ?(w)1/2 ~ Var(vT(w?1/2 when
?(w) ~ O-while Blumer et al. (1989) and Anthony and Biggs (1992) work with
P(suPWEW IVT(W) - ?(w)ll{o}(VT(W? ~ i). The latter obtain the sufficient condition
(7)
nc = (5.8v/i) log 12/i
for nets, if any, having VT( w) = o. If one is guaranteed to do reasonably well on
the training set, a smaller order of dependence results.
Results (Turmon & Fine, 1993) for perceptrons and P a Gaussian mixture imply
that at least v/280i 2 samples are needed to force ?(w?) - ?(WO) < 2i with high
probability. (Here w? is the best linear discriminant with weights estimated from
the data.) Combining with Talagrand's result, we see that the general (not assuming
small VT(W? functional dependence is V/i 2.
Sample Size Requirements for Feedforward Neural Networks
2
329
APPLYING THE POISSON CLUMPING HEURISTIC
We adopt a new approach to the problem. For the moderately large values of n
we anticipate, the central limit theorem informs us that Vn[lIT(W) - E(w)] has
nearly the distribution of a zero-mean Gaussian random variable. It is therefore
reasonable l to suppose that
P( sup IlIT(W) - E(w)1 ~ f) ~ P( sup IZ(w)1 ~ fJ1i) ~ 2P( sup Z(w) ~ fVn)
wEW
wEW
wEW
where Z( w) is a Gaussian process with mean zero and covariance
R(w, v)
= EZ(w)Z(v) = Cov(y -1J(x; w?2, (y -1J(x; V?2)
The problem about extrema of the original empirical process is equivalent to one
about extrema of a corresponding Gaussian process.
The Poisson clumping heuristic (PCR), introduced in the remarkable (Aldous,
1989), provides a general tool for estimating such exceedance probabilities. Consider the excursions above level b(= fVn ~ 1) by a stochastic process Z(w). At
left below, the set {w : Z( w) ~ b} is seen as a group of "clumps" scattered in weight
space W. The PCR says that, provided Z has no long-range dependence and the
level b is large, the centers of the clumps fall according to the points of a Poisson
process on W, and the clump shapes are independent. The vertical arrows (below
right) illustrate two clump centers (points of the Poisson process); the clumps are
the bars centered about the arrows.
w
w
In fact, with PheW) = P(Z(w) ~ b), Ch(W) the size of a clump located at w, and
Ah (w) the rate of occurrence of clump centers, the fundamental equation is
(8)
The number of clumps in W is a Poisson random variable Nh with parameter
The probability of a clump is P(Nb > 0) = 1- exp( - fwAh( w) dW) =:::
Ah(W) dw where the approximation holds because our goal is to operate in a
regime where this probability is near zero. Letting ~(b) = P(N(0, 1) > b) and
(T2(w) = R(w, w), we have PheW) = ~(b/(T(w?. The fundamental equation becomes
1, Ah( w) dw.
fw
P( sup Z(w)
wEW
~ b) ~
r ~(b/(T(w?
ECh(W)
Jw
dw
(9)
It remains only to find the mean clump size ECh( w) in terms of the network architecture and the statistics of (x, y).
lSee ch. 7 of (Pollard, 1984) for treatment of some technical details in this limit.
330
3
Michael Tunnon, Terrence L. Fine
POISSON CLUMPING FOR SMOOTH PROCESSES
Assume Z(w) has two mean-square derivatives in w. (If the network activation
functions have two derivatives in w, for example, Z( w) will have two almost sure
derivatives.) Z then has a parabolic approximation about some Wo via its gradient
G = 'VZ(w) and Hessian matrix H = 'V'VZ(w) at woo Provided Zo ~ b, that is
that there is a clump at Wo, simple computations reveal
Cb( wo) ~ Kd
(2(Zo - b) - cP'H- 1 G)d/2
IHI I / 2
where Kd is the volume of the unit ball in Rd and I? 1is the determinant.
clump size is the expectation of this conditioned on Z(wo)
~
(10)
The mean
b.
The same argument used to show that Z(w) is approximately normal shows that G
and H are approximately normal too. In fact,
E[HIZ(wo)
z
(F 2( Wo )A(wo)
-EZ(wo)H = -'Vw'VwR(wo, w)lw=wo
= z]
A(wo)
so that, since b (and hence z) is large, the second term in the numerator of (10)
may be neglected. The expectation is then easily computed, resulting in
Lemma 1 (Smooth process clump size) Let the network activation functions
be twice continuously differentiable, and let b ? (F( w). Then
ECb(W)
~ (21r)d/21 ~~~) 1- 112 ((F(:?) d
Substituting into (9) yields
P( sup Z(w)
~ b) ~ (21r)-~
1A(w) 11/2 (_b_) d-~_b~/2q~(W) dw,
(11)
(F(w)
where use of the asymptotic expansion ~(z) ~ (zv'21r)-l exp( _Z2 /2) is justified
since ('v'w)b ? (F( w) is necessary to have the individual P( Z( w) ~ b) low-let alone
the supremum. To go farther, we need information about the variance (F2 (w) of
(y - 11( x; w?2. In general this must come from the problem at hand, but suppose
wEW
(
iw
(F2(w)
for example the process has a unique variance maximum 0'2 at w. Then, since
the level b is large, we can use Laplace's method to approximate the d-dimensional
integral.
Laplace's method finds asymptotic expansions for integrals
fw g(w) exp( - f(w)2 /2) dw
when few) is C2 with a unique positive minimum at Wo in the interior of W ~ Rd ,
and g( w) is positive and continuous. Suppose I( wo) ? 1 so that the exponential
factor is decreasing much faster than the slowly varying g. Expanding f to second
order about Wo, substituting into the exponential, and performing the integral shows
that
iw
g( w) exp( - f(w)2 /2) dw
~ (21r)d/2If( wo)KI- 1/ 2g( wo) exp( -
f( wo)2 /2)
Sample Size Requirements for Feedforward Neural Networks
331
where K = V'V'f(w)lwo, the Hessian of f. See (Wong, 1989) for a proof. Applying
this to (11) and using the asymptotic expansion for ~ in reverse yields
Theorem 1 Let the network activation functions be twice continuously differentiable. Let the variance have a unique maximum u at w in the interior of Wand
the level b ~ u. Then the peH estimate of exceedance probability is given by
IA(w)1 1/ 2
_
P(:~fv Z(w) ~ b) ~ IA(w) _ r(w)1 1/ 2 ~(b/u)
(12)
where r(w) = V'wV'tlR(w,v)lw=tI=w. Furthermore, A- r is positive-definite at w;
it is -1/2 the Hessian of cr 2 (w). The leading constant thus strictly e:cceeds unity.
The above probability is just P(Z(w) ~ b) multiplied by a factor accounting for the
other networks in the supremum. Letting b = f...;n reveals
nc
=
u2 10g(IA(w)I/IA(w) - r(w)!)
f. 2
(13)
samples force P(supw IlIT(W) - &(w)\ ~ {) below unity. If the variance maximum is
not unique but occurs over a d-dimensional set within W, the sample size estimate
becomes proportional to u2d/{2. With d playing the role of VC dimension v, this
is similar to Vapnik's bound although we retain dependence on P and N.
The above probability is determined by behavior near the maximum-variance point,
which for example in classification is where &(w) = 1/2. Such nets are uninteresting as classifiers, and certainly it is undesirable for them to dominate the entire
probability. This problem is avoided by replacing Z(w) with Z(w)/cr(w), which additionally allows a finer resolution where &(w) nears zero. Indeed, for classification,
if n is such that with high probability
sup
weW
IlIT(W) - &(w)1
IlIT(W) - &(w)1
= sup
<{ ,
cr(w)
wew J&(W)(I- &(w?
(14)
then lIT(W?) = 0 ::} &(w?) < {2(1 + (2)-1 ~ {2 <t:: {. Near lIT(W?) = 0, condition (14)/ is much more powerful than the corresponding unnormalized one. Sample
size estimates using this setup give results having a functional form similar to (7).
4
ANOTHER MEANS OF COMPUTING CLUMP SIZE
Conditional on there being a clump center at w, the upper bound
Cb(W)
~ Db(W) ==
iw
l[o,oo)(Z(w') - b) dw'
(15)
is evidently valid: the volume of the clump at w is no larger than the total volume of
all clumps. (The right hand side is indeed a function of w because we condition on
occurrence of a clump center at w.) The bound is an overestimate when the number
Nb of clumps exceeds one, but recall that we are in a regime where b (equivalently
n) is large enough so that P( Nb > 1)/ P( Nb = 1) ~ fw ).b (w) dw <t:: 1. Thus error
in (15) due to this source is negligible. To compute its mean, we approximate
EDb(W)
=
iw P(Z(~f) ~
blw a clump center)dw'
332
Michael Turmon, Terrence L. Fine
(16)
The point is that occurrence of a clump center at Wo is a smaller class of events than
merely Z( wo) ~ b: the latter can arise from a clump center at a nearby w E W
capturing woo Since Z(w) and Z(w') are jointly normal, abbreviate u = u(w),
u' = u(w'), p = p(w,w') = R(w,w')/(uu'), and let
(=(w,w')
(U I U,)I;r;!;
1- p2
(1- p)/(1 + p?)1/2
(17)
(constant variance case)
(18)
Evaluating the conditional probabilities of (16) presents no problem, and we obtain
Lemma 2 (Clump size estimate) For b ~ u the mean clump size is
ECb(W)
~ EDb(W) ~
/w cf>?blu)() dw'
(19)
Remark 1. This integral will be used in (9) to find
P(s~pZ(w) > b) ~
(
cf>(blu)
Jw fw~?blu)() dw,dw
(20)
Since b is large, the main contribution to the outer integral occurs for w near a
variance maximum, i.e. for u' I u ~ 1. If the variance is constant then all w E W
contribute. In either case ( is nonnegative. By lemma 1 we expect (19) to be, as
a function of b, of the form (const ulb)P for, say, p = d. In particular, we do not
anticipate the exponentially small clump sizes resulting if (Vw')( w, w') ~ M ~ O.
Therefore ( should approach zero over some range of w', which happens only when
p ~ 1, that is, for w' near w. The behavior of pew, w') for w' ~ w is the key to
finding the clump size.
Remark 2. There is a simple interpretation of the clump size; it represents the
volume of w' E W for which Z(w') is highly correlated with Z(w) . The exceedance
probability is a sum of the point exceedance probabilities (the numerator of (20?,
each weighted according to how many other points are correlated with it. In effect,
the space W is partitioned into regions that tend to "have exceedances together,"
with a large clump size ECb( w) indicating a large region. The overall probability
can be viewed as a sum over all these regions of the corresponding point exceedance
probability. This has a similarity to the Vapnik argument which lumps networks
together according to their nV Iv! possible actions on n items in the training set. In
this sense the mean clump size is a fundamental quantity expressing the ability of
an architecture to generalize.
5
EMPIRICAL ESTIMATES OF CLUMP SIZE
The clump size estimate of lemma 2 is useful in its own right if one has information
about the covariance of Z. Other known techniques of finding ECb( w) exploit
special features of the process at hand (e.g. smoothness or similarity to other wellstudied processes); the above expression is valid for any covariance structure. In
Sample Size Requirements for Feedf01ward Neural Networks
333
this section we show how one may estimate the clump size using the training set,
and thus obtain probability approximations in the absence of analytical information
about the unknown P and the potentially complex network architecture N.
Here is a practical way to approximate the integral giving EDb{W). For'Y
a set of significant w'
S-y{W) = {w' E W: (w,w') $ 'Y}
then monotonicity of ~ yields EDb{W) ~
< 1 define
V-y{W) = vol{S-y(w))
(21)
Is .., ~((b/(1X) dw' ~ V-y(W) ~((b/uh) .
This apparently crude lower bound for ~ is accurate enough near the origin to give
satisfactory results in the cases we have studied. For example, we can characterize
the covariance R( w, w') of the smooth process oflemma 1 and thus find its ( function. The bound above is then easily calculated and differs by only small constant
factors from the clump size in the lemma.
The lower bound for EDb(W) yields the upper bound
1
~(b/(1-)
P(s~p Z(w) ~ b) $ w V-y(w) ~?b/uh) dw
(22)
We call V-y(w) the correlation volume, as it represents those weight vectors w' whose
errors Z(w') are highly correlated with Z(w); one simple way to estimate the correlation volume is as follows . Select a weight w' and using the training set compute
(Yl - 17( Xl; w))2, ... , (Yn - 17( Xn; w)? & (Yl - 17( Xl; w'))2 , ... , (Yn - 17( Xn; w'))2 .
It is then easy to estimate u 2, u,2, and p, and finally (w , w'), which is compared
to the chosen 'Y to decide if w' E S-y (w) .
The difficulty is that for large d, S-y (w) is far smaller than any approximatelyenclosing set. Simple Monte Carlo sampling and even importance sampling methods
fail to estimate the volume of such high-dimensional convex bodies because so few
hits occur in probing the space (Lovasz, 1991). The simplest way to concentrate the
search is to let w' = w except in one coordinate and probe along each coordinate
axis. The correlation volume is approximated as the product of the one-dimensional
measurements.
Simulation studies of the above approach have been performed for a perceptron
architecture in input uniform over [-1, l]d. The integral (22) is computed by Monte
Carlo sampling, and based ona training set of size lOOd, V-y (w) is computed at each
point via the above method. The result is that an estimated sample size of 5.4d/f 2
is enough to ensure (14) with high probability. For nets, if any, having VT(W) = 0,
sample sizes larger than 5.4d/f will ensure reliable generalization, which compares
favorably with (7) .
6
SUMMARY AND CONCLUSIONS
To find realistic estimates of sample size we transform the original problem into
one of finding the distribution of the supremum of a derived Gaussian random
field, which is defined over the weight space of the network architecture. The
latter problem is amenable to solution via the Poisson clumping heuristic. In terms
of the PCH the question becomes one of estimating the mean clump size, that
334
Michael Turman, Terrence L. Fine
is, the typical volume of an excursion above a given level by the random field.
In the "smooth" case we directly find the clump volume and obtain estimates of
sample size that are (correctly) of order v/?2. The leading constant, while explicit,
depends on properties of the architecture and the data-which has the advantage
of being tailored to the given problem but the potential disadvantage of our having
to compute them.
We also obtain a useful estimate for the clump size of a general process in terms of
the correlation volume V-y(w). For normalized error, (22) becomes approximately
p (sup lIr(w) - ?(w)
weW
u(w)
> ?) ~ E
-
[vol(W)]
e-(1--y2)nf 2 /2
V-y(w)
where the expectation is taken with respect to a uniform distribution on W. The
probability of reliable generalization is roughly given by an exponentially decreasing
factor (the exceedance probability for a single point) times a number representing
degrees of freedom. The latter is the mean size of an equivalence class of "similarlyacting" networks. The parallel with the Vapnik approach, in which a worst-case
exceedance probability is multiplied by a growth function bounding the number of
classes of networks in N that can act differently on n pieces of data, is striking. In
this fashion the correlation volume is an analog of the VC dimension, but one that
depends on the interaction of the data and the architecture.
Lastly, we have proposed practical methods of estimating the correlation volume
empirically from the training data. Initial simulation studies based on a perceptron
with input uniform on a region in Rd show that these approximations can indeed
yield informative estimates of sample complexity.
References
Aldous, D. 1989. Probability Approximations via the Poisson Clumping Heuristic.
Springer.
Anthony, M., & Biggs, N. 1992. Computational Learning Theory. Cambridge Univ.
Baum, E., & Haussler, D. 1989. What size net gives valid generalization? Pages
81-90 of' Touretzky, D. S. (ed), NIPS 1.
Blumer, A., Ehrenfeucht, A., Haussler, D., & Warmuth, M. K. 1989. Learnability
and the Vapnik-Chervonenkis dimension. Jour. Assoc. Compo Mach., 36,929-965.
LovMz, L. 1991. Geometric Algorithms and Algorithmic Geometry. In: Proc.
Internat. Congr. Mathematicians. The Math. Soc. of Japan.
Pollard, D. 1984. Convergence of Stochastic Processes. Springer.
Talagrand, M. 1994. Sharper bounds for Gaussian and empirical processes. Ann.
Probab., 22, 28-76.
Turmon, M. J., & Fine, T. L. 1993. Sample Size Requirements of Feedforward
Neural Network Classifiers. In: IEEE 1993 Intern. Sympos. Inform. Theory.
Vapnik, V. 1982. Estimation of Dependences Based on Empirical Data. Springer.
Wong, R. 1989. Asymptotic Approximations of Integrals. Academic.
| 970 |@word determinant:1 blu:3 ona:1 seek:1 simulation:2 covariance:4 accounting:1 initial:1 chervonenkis:2 existing:1 z2:1 comell:2 activation:3 must:2 realistic:1 informative:1 shape:1 drop:1 clumping:6 alone:1 item:1 warmuth:1 farther:1 compo:1 provides:1 math:1 contribute:1 shatter:1 c2:1 along:1 sustained:1 indeed:3 roughly:1 behavior:2 decreasing:2 becomes:4 provided:2 estimating:3 what:1 mathematician:1 extremum:2 finding:3 nf:1 ti:1 act:1 growth:1 classifier:5 k2:1 hit:1 assoc:1 unit:1 yn:2 overestimate:1 positive:3 negligible:1 engineering:2 limit:2 mach:1 approximately:3 twice:2 studied:1 equivalence:1 limited:1 range:2 clump:36 ihi:1 unique:5 practical:3 practice:2 definite:1 differs:1 empirical:5 interior:2 undesirable:1 nb:4 applying:2 wong:2 equivalent:1 center:8 baum:2 go:1 convex:1 restated:1 resolution:2 haussler:3 dominate:1 dw:15 lsee:1 coordinate:2 laplace:2 suppose:3 origin:1 approximated:1 located:1 role:1 electrical:2 worst:1 region:4 ilit:4 transforming:1 inaccessible:1 complexity:3 moderately:1 neglected:1 f2:2 supwew:2 basis:1 biggs:2 uh:2 easily:2 differently:1 train:1 univ:3 zo:2 ech:2 monte:2 sympos:1 whose:2 heuristic:5 larger:2 turmon:4 say:2 ability:1 cov:1 statistic:1 jointly:1 transform:1 advantage:1 differentiable:2 evidently:1 net:9 analytical:1 interaction:1 product:1 combining:1 convergence:1 requirement:5 illustrate:1 informs:1 oo:1 measured:1 lwo:1 p2:1 soc:1 involves:1 come:1 uu:1 concentrate:1 stochastic:2 vc:3 centered:1 generalization:3 anticipate:2 strictly:1 hold:2 normal:3 exp:5 fvn:2 cb:2 algorithmic:1 substituting:2 adopt:1 purpose:1 ecb:4 estimation:1 proc:1 iw:4 largest:1 vz:2 peh:1 tool:1 weighted:1 lovasz:1 gaussian:6 avoid:1 cr:3 cornell:2 varying:1 derived:1 indicates:1 sense:1 nears:1 entire:1 issue:1 among:1 supw:1 classification:2 overall:1 raised:1 special:1 edb:5 field:3 having:4 reasoned:1 sampling:3 represents:2 lit:3 nearly:1 t2:1 employ:1 few:2 individual:1 geometry:1 freedom:1 investigate:1 highly:2 certainly:1 wellstudied:1 mixture:1 yielding:1 amenable:1 accurate:1 integral:8 necessary:1 experience:1 iv:1 desired:1 classify:1 disadvantage:1 lood:1 uninteresting:1 uniform:3 too:1 learnability:1 characterize:1 jour:1 fundamental:3 retain:1 terrence:5 yl:2 michael:5 together:2 continuously:2 squared:1 central:1 slowly:1 exceedance:8 mjt:1 stochastically:1 derivative:3 leading:2 japan:1 potential:1 depends:2 piece:1 performed:1 apparently:1 sup:11 parallel:1 wew:11 contribution:1 square:1 variance:8 who:1 yield:6 ofthe:1 generalize:1 carlo:2 finer:2 ah:3 inform:1 touretzky:1 ed:1 proof:1 treatment:1 recall:1 higher:1 jw:2 formulation:1 furthermore:1 just:1 lastly:1 correlation:6 talagrand:3 hand:3 replacing:1 reveal:1 perhaps:1 effect:1 normalized:1 y2:1 hence:1 satisfactory:1 ehrenfeucht:1 ll:1 numerator:2 unnormalized:1 cp:1 consideration:1 recently:1 functional:2 empirically:1 exponentially:3 nh:1 volume:13 analog:1 interpretation:1 expressing:1 significant:1 measurement:1 cambridge:1 pew:1 rd:4 smoothness:1 similarity:2 internat:1 own:2 recent:1 aldous:2 reverse:1 wv:1 vt:7 yi:1 seen:1 minimum:1 smooth:4 technical:1 match:1 academic:1 faster:1 exceeds:1 long:1 expectation:3 poisson:9 tailored:1 justified:1 fine:7 source:1 ithaca:2 operate:1 sure:1 nv:1 tend:1 db:1 lump:1 practitioner:1 call:1 ee:2 near:7 noting:1 vw:2 feedforward:5 enough:4 easy:1 architecture:11 expression:1 forecasting:1 wo:25 pollard:2 hessian:3 remark:2 action:1 useful:3 simplest:1 estimated:2 correctly:1 ivt:7 iz:1 vol:2 group:1 zv:1 key:1 merely:1 sum:2 wand:1 parameterized:1 powerful:1 striking:1 almost:1 reasonable:1 parabolic:1 decide:1 vn:1 excursion:2 capturing:1 bound:13 ki:1 guaranteed:1 tlr:1 nonnegative:1 occur:1 nearby:1 argument:2 performing:1 according:4 ball:1 kd:2 smaller:3 unity:3 partitioned:1 happens:1 restricted:1 taken:1 equation:2 remains:1 turn:1 fail:1 needed:1 letting:2 b_:1 multiplied:2 probe:1 occurrence:3 original:2 ensure:4 cf:2 hiz:1 const:1 exploit:1 giving:1 question:1 quantity:2 occurs:2 ulb:1 dependence:5 gradient:1 outer:1 discriminant:1 fresh:1 assuming:1 nc:4 setup:1 equivalently:1 potentially:1 sharper:1 favorably:1 unknown:3 upper:2 vertical:1 attacked:1 community:1 introduced:2 required:1 kl:2 sentence:1 fv:1 narrow:1 nip:1 bar:1 below:4 regime:2 pcr:2 reliable:2 ia:4 event:1 difficulty:1 force:3 abbreviate:1 representing:1 imply:1 axis:1 woo:2 probab:1 geometric:1 disagrees:1 determining:1 asymptotic:4 law:1 expect:1 gauged:1 proportional:1 var:1 remarkable:2 degree:1 sufficient:2 playing:1 summary:1 last:1 side:1 allow:1 perceptron:2 fall:1 face:1 taking:1 dimension:4 calculated:1 valid:3 evaluating:1 xn:2 avoided:1 far:1 approximate:3 supremum:4 dealing:1 monotonicity:1 lir:1 reveals:1 xi:2 continuous:1 search:1 additionally:1 reasonably:1 expanding:1 expansion:3 complex:1 anthony:2 main:1 arrow:2 bounding:1 arise:1 body:1 en:1 scattered:1 fashion:1 ny:2 probing:1 explicit:1 exponential:2 xl:2 crude:1 lw:2 rk:1 theorem:2 pz:1 vapnik:8 importance:1 magnitude:2 conditioned:1 gap:1 intern:1 ez:2 u2:1 springer:3 ch:2 conditional:2 goal:1 viewed:1 blumer:2 ann:1 absence:1 fw:4 determined:1 except:1 typical:1 lemma:5 called:1 total:1 perceptrons:1 indicating:1 select:2 latter:4 correlated:3 |
7,211 | 971 | Combining Estimators Using
Non-Constant Weighting Functions
Volker Tresp*and Michiaki Taniguchi
Siemens AG, Central Research
Otto-Hahn-Ring 6
81730 Miinchen, Germany
Abstract
This paper discusses the linearly weighted combination of estimators in which the weighting functions are dependent on the input .
We show that the weighting functions can be derived either by
evaluating the input dependent variance of each estimator or by
estimating how likely it is that a given estimator has seen data in
the region of the input space close to the input pattern. The latter solution is closely related to the mixture of experts approach
and we show how learning rules for the mixture of experts can be
derived from the theory about learning with missing features. The
presented approaches are modular since the weighting functions
can easily be modified (no retraining) if more estimators are added. Furthermore, it is easy to incorporate estimators which were
not derived from data such as expert systems or algorithms.
1
Introduction
Instead of modeling the global dependency between input x E ~D and output y E ~
using a single estimator, it is often very useful to decompose a complex mapping
-'\.t the time of the research for this paper, a visiting researcher at the Center for
Biological and Computational Learning, MIT. Volker.Tresp@zfe.siemens.de
420
Volker Tresp, Michiaki Taniguchi
into simpler mappings in the form l
(1)
M
n(x) =
L hi(X)
i=l
The weighting functions hi(X) act as soft switches for the modules N Ni(X). In the
mixture of experts (Jacobs et al., 1991) the decomposition is learned in an unsupervised manner driven by the training data and the main goal is a system which learns
quickly. In other cases, the individual modules are trained individually and then
combined using Equation 1. We can distinguish two motivations: first, in the work
on averaging estimators (Perrone, 1993, Meir , 1994, Breiman, 1992) the modules
are trained using identical data and the weighting functions are constant and, in
the simplest case, all equal to one. The goal is to achieve improved estimates by
averaging the errors of the individual modules. Second, a decomposition as described in Equation 1 might represent some "natural" decomposition of the problem
leading to more efficient representation and training (Hampshire and Waibel, 1989).
A good example is a decomposition into analysis and action. hi(x) might be the
probability of disease i given the symptoms x, the latter consisting of a few dozen
variables. The amount of medication the patient should take given disease i on the
other hand - represented by the output of module N Ni (x) - might only depend
on a few inputs such as weight, gender and age.2 Similarly, we might consider hie x)
as the IF-part of the rule, evaluating the weight of the rule given x, and as N Ni(X)
the conclusion or action which should be taken under rule i (compare Tresp, Hollatz and Ahmad, 1993). Equation 1 might also be the basis for biological models
considering for example the role of neural modulators in the brain. Nowlan and
Sejnowsky (1994) recently presented a biologically motivated filter selection model
for visual motion in which modules provide estimates of the direction and amount
of motion and weighting functions select the most reliable module.
In this paper we describe novel ways of designing the weighting functions. Intuitively, the weighting functions should represent the competence or the certainty of
a module, given the available information x. One possible measure is related to the
number of training data that a module has seen in the neighborhood of x. Therefore, P(xli), which is an estimate of the distribution of the input data which were
used to train module i is an obvious candidate as weighting function. Alternatively,
the certainty a module assigns to its own prediction, represented by the inverse of
the variance 1/ var( N Ni ( x? is a plausible candidate for a weighting function. Both
approaches seem to be the flip-sides of the same coin, and indeed, we can show that
both approaches are extremes of a unified approach.
IThe hat stands for an estimates value.
2Note, that we include the case that the weighting functions and the modules might
explicitly only depend on different subsets of x.
421
Combining Estimators Using Non-Constant Weighting Functions
(e)
1
0.8
I,
0.6
I
/ /
\
I
\
\
\
0 .4
0.2
I
\
(d)
(b)
0.5r-r-----T"I
I '
I,:'
I:
j
,l
t'
(f)
0.5.----------"
o
:1
~
-1
-1
-1
-1
o
-1
o
1
o
-1
1
Figure 1: (a): Two data sets (1:*, 2:0) and the underlying function (continuous).
(b) The approximations of the two neural networks trained on the data sets (continuous: 1, dashed: 2). Note, that the approximation of a network is only reliable in
the regions of the input space in which it has "seen" data. (c) The weighting functions for variance-based weighting. (d) The approximation using variance-based
weighting (continuous). The approximation is excellent, except to the very right.
(e) The weighting functions for density-based weighting (Gaussian mixtures approximation). (f) The approximation using density-based weighting (continuous). In
particular to the right, the extrapolation is better than in (d).
2
Variance-based Weighting
Here, we assume that the different modules N Ni(x) were trained with different
data sets {(xLY~)}:~l but that they model identical input-output relationships
(see Figure 1 a,b ). To give a concrete example, this would correspond to the case
that we trained two handwritten digit classifiers using different data sets and we
want to use both for classifying new data.
If the errors of the individual modules are uncorrelated and unbiased, 3 the combined
estimator is also unbiased and has the smallest variance if we select the weighting
functions inversely proportional the the variance of the modules
1
(2)
hi(X) = var(NNi(x))
This can be shown using var(2:::'l gi(x)N Ni(X)) = 2:::'1 gl(x)var(N Ni(X)) and
1. Intuitively,
using Lagrange mUltiplier to enforce the constraint that 2:i gi(X)
=
3The errors are un correlated since the modules were trained with different data; correlation and bias are discussed in Section 8.1.
422
Volker Tresp, Michiaki Taniguchi
Equation 2 says that a module which is uncertain about its own prediction should
also obtain a smaller weight. We estimate the variance from the training data as
var
? '" 8N Ni(xl H-:- 18N Ni(X)
* x ""
8w
*
8w'
(NN .(
Hi is the Hessian, which can be approximated as
Tibshirani, 1994)
3
?(12 is the output-noise variance,
Density-based Weighting
In particular if the different modules were trained with data sets from different
regions of the input space, it might be a reasonable assumption that the different
modules represent different input-output relationships. In terms of our example,
this corresponds to the problem, that we have two handwritten digit classifiers, one
trained with American data and one with European data. If the classifiers are used
in an international setting, confusions are possible, since, for example, an American
seven might be confused with a European one. Formally, we introduce an additional
variable which is equal to zero if the writer is American and is equal to one if the
writer is European. During recall, we don't know the state of that variable and
we are formally faced with the problem of estimation with missing inputs. From
previous work (Ahmad and Tresp, 1993) we know that we have to integrate over
the unknown input weighted by the conditional probability of the unknown input
given the known variables. In this case, this translates into Equation 1, where the
weighting function is
In our example, P(ilx) would estimate the probability that the writer is American
or European given the data.
Depending on the problem P(ilx) might be estimated in different ways. If x represents continuous variables, we use a mixture of Gaussians model
(3)
where G(x; cij, Eij) is our notation for a normal density centered at cij and with
covariance Eij.
Note that we have obtained a mixture of experts network with P( ilx) as gating network. A novel feature of our approach is that we maintain an estimate of the input
data distribution (Equation 3), which is not modeled in the original mixture of experts network. This is advantageous if we have training data which are not assigned
423
Combining Estimators Using Non-Constant Weighting Functions
to a module (in the mixture of experts, no data are assigned) which corresponds
to training with missing inputs (the missing input is the missing assignment), for
which the solution is known (Tresp et ai., 1994). If we use Gaussian mixtures to
approximate P(xli), we can use generalized EM learning rules for adaptation. The
adaptation of the parameters in the "gating network" which models P(xli) is therefore somewhat simpler than in the original mixture of experts learning rules (see
Section 8.2).
4
Unified Approach
In reality, the modules will often represent different mappings, but these mappings
are not completely independent. Let's assume that we have an excellent American
handwritten digit classifier but our European handwritten digit classifier is still very
poor, since we only had few training data. We might want to take into account the
results of the American classifier, even if we know that the writer was European.
Mathematically, we can introduce a coupling between the modules. Let's assume
that the prediction ofthe i-th module NNi(X) = li(x)+{i is a noisy version of the
true underlying relationship li(x) and that {i is independent Gaussian noise with
variance var(N Ni(X?. Furthermore, we assume that the true underlying functions
are coupled through a prior distribution (for simplicity we only assume two modules)
1
P(h(x), h(x)) ex: exp (--2-(I1(x)
- 12(x ) )2 ).
vare
We obtain as best estimates
A
1
h(x) = K(x) [(var(N N 2 (x))
A
1
+ vare) N Nt(x) + var(N Nl(x))
h(x) = K(x) [var(N N2(x? N Nl(X) + (var(N Nt (x?
where
K(x)
+ vare)
N N2(X)]
N N2(X)]
= var(N Nl (x? + var(N N2 (x)) + vare.
=
We use density-based weighting to combine the two estimates: y(x) P(llx)it(x)+
P(2Ix)i2(X). Note, that if vare -- 00 (no coupling) we obtain the density-based
solution and for vare -- 0 (the mappings are forced to be identical) we obtain the
variance-based solution. A generalization to more complex couplings can be found
in Section 8.2.1.
5
Experiments
We tested our approaches using the Boston housing data set (13 inputs, one continuous output). The training data set consisted of 170 samples which were divided
into 20 groups using k-means clustering. The clusters were then divided randomly
into two groups and two multi-layer perceptrons (MLP) were trained using those two
424
Volker Tresp. Michiaki Taniguchi
data sets. Table 1 shows that the performances of the individual networks are pretty
bad which indicates that both networks have only acquired local knowledge with
only limited extrapolation capability. Variance-based weighting gives considerably
better performance, although density-based weighting and the unified approach are
both slightly better. Considering the assumptions, variance-based weighting should
be superior since the underlying mappings are identical. One problem might be
that we assumed that the modules are unbiased which might not be true in regions
were a given module has seen no data.
Table 1: Generalization errors
N N2
0.6948
6
1.188
I variance-based I density-based I unified I
I 0.4821
I 0.4472
I 0.4235 I
Error-based Weighting
In most learning tasks only one data set is given and the task is to obtain optimal
predictions. Perrone (1994) has shown that simply averaging the estimates of a
small number (i. e. 10) of neural network estimators trained on the same training data set often gives better performance than the best estimator out of this
ensemble. Alternatively, bootstrap samples of the original data set can be used for
training (Breimann, personal communication). Instead of averaging, we propose
that Equation 1, where
might give superior results (error-based weighting). Res(N Ni(X)) stands for an
estimate of the input dependent residual squared error at x. As a simple approximation, Res(N Ni(X)) can be estimated by training a neural network with the
residual squared errors of N Ni. Error-based weighting should be superior to simple averaging in particular if the estimators in the pool have different complexity.
A more complex system would obtain larger weights in regions where the mapping is complex, since an estimator which is locally too simple has a large residual
error, whereas in regions, where the mapping is simple, both estimators have sufficient complexity, but the simpler one has less variance. In our experiments we
only tried networks with the same complexity. Preliminary results indicate that
variance-based weighting and error-based weighting are sometimes superior to simple averaging. The main reason seems to be that the local overfitting of a network
is reflected in a large variance near that location in input space. The overfitting
estimator therefore obtains a small weight in that region (compare the overfitting
of network 1 in Figure Ib near x
0 and the small weight of network 1 close to
x 0 in Figure lc).
=
=
425
Combining Estimators Using Non-Constant Weighting Functions
7
Conclusions
We have presented modular ways for combining estimators. The weighting functions
of each module can be determined independently of the other modules such that
additional modules can be added without retraining of the previous system. This
can be a useful feature in the context of the problem of catastrophic forgetting:
additional data can be used to train an additional module and the knowledge in the
remaining modules is preserved. Also note that estimators which are not derived
from data can be easily included if it is possible to estimate the input dependent
certainty or competence of that estimator.
Acknowledgements: Valuable discussions with David Cohn, Michael Duff and
Cesare Alippi are greatfully acknowledged. The first author would like to thank
the Center for Biological and Computational Learning (MIT) for providing and
excellent research environment during the summer of 1994.
8
8.1
Appendix
Variance-based Weighting: Correlated Errors and Bias
We maintain that Li gi(X) = 1. In general (Le. the modules have seen the same data, or
partially the same data), we cannot assume that the errors in the individual modules are
independent. Let the M x M matrix O( x) be the covariance between the predictions of the
modules N Ni(X). With h(x) (h 1(x) .... hM(X)T, the optimal weighting vector becomes
=
h(x) = 0-1(X)
n(x)
U
= u' 0-1(X)
U
where u is the M-dimensional unit vector.
If the individual modules are biased (biasi(x) = ED(N Ni(X)) - EYI.r(ylx)),~ we form the
M x M matrix B(x), with Bii(x) = biasi(x)biasj(x), and the minimum variance solution
is found for
h(x) = (O(x)
8.2
+ B(X))-1 u
n(x) = u' (O(x)
+ B(X))-1 u.
Density-based Weighting: GEM-learning
Let's assume a training pattern (Xlo' YIo) which is not associated with a particular module.
If wi is a parameter in network N Ni the error gradient becomes
8errorlo __ (Ylo -NNo
, ( Xlo ))p"(OIS Xlo, Ylo )8NNi(XIo)
8
8 w'
w'
?
-
?
?
This equation can be derived from the solution to the problem of training with missing
features (here: the true i is unknown, see Tresp, Ahmad and Neuneier, 1994). This
corresponds also to the M-step in a generalized EM algorithm, where the E-step calculates
" .
P(SIXIo, YIo)
P(Ylolxlo , i)P(xloli)P(i)
= '"
p
o"
L..Ji (Ylolxlo, s)P(xlols)P(s)
0
"
.
"
?
P(Ylolxlo, s)
= G(ylo; N Ni(XIo) ,
2
(1
).
~ E stands for the expected value; the expectation ED is taken with respect to all data
sets of the same size.
426
Volker Tresp, Michiaki Taniguchi
using the current parameters. The M-step in the "gating network" P(xli) is particularly
simple using the well known EM-rules for Gaussian mixtures. Note, that P(module =
i, mixture component: ilxk' Yk) needs to be calculated.
8.2.1
Unified Approach: Correlated Errors and General Coupling
Let's form the vectors N N(x) = (N N 1(x), ... N NM(X))T and /(x) = (!t(x), ... , /M(x)f.
In a more general case, the prior coupling between the underlying functions is described
by
P(f(x)) = G(f(x);g(x), ~g(x))
=
where g{x)
(g1{X), ... ,gM{x)f. Furthermore, in a more general case, the estimates are
not independent,
P{N N{x)l/{x)) = G{N N(x); I(x), ~N{X)).
The minimum variance solution is now
=
The equations in Section 4 are special cases with M = 2, g{x) = 0, ~;1(x)
l/varcc x
(1, -1)(1, -If, ~N(X) = 1 (var{N N 1{x)), var(N N2{X)))T (1 is the 2D-unit matrix).
References
Ahmad, S. and Tresp, V. (1993) . Some Solutions to the Missing Feature Problem in Vision.
In S. J. Hanson, J. D. Cowan and C. L. Giles, (Eds.), Advances in Neural Information
Processing Systems 5. San Mateo, CA: Morgan Kaufmann.
Breiman, L. (1992). Stacked Regression. Dept. of Statistics, Berkeley, TR No. 367.
Hampshire, J. and Waibel, A. (1989). The meta-pi network: Building Distributed Knowledge Representations for Robust Pattern Recognition. TR CMU-CS-89-166, CMU, PA.
Jacobs, R. A., Jordan, M. 1., Nowlan, S. J. and Hinton, J. E. (1991). Adaptive Mixtures
of Local Experts. Neuml Computation, Vol. 3, pp. 79-87.
Meir, R. (1994). Bias, Variance and the Combination of Estimators: The Case of Linear
Least Squares. TR: Dept. of Electrical Engineering, Technion, Haifa.
Nowlan, S. J and Sejnowski, T. J. (1994). Filter Selection Model for Motion Segmentation
and Velocity Integration. J. Opt. Soc. Am. A, Vol. 11, No. 12, pp. 1-24.
Perrone, M. P. (1993). Improving Regression Estimates: Averaging Methods for Variance
Reduction with Extensions to General Convex Measure Optimization. PhD thesis. Brown
University.
Tibshirani, R. (1994). A Comparison of Some Error Estimates for Neural Network Models.
TR Department of Statistics, University of Toronto.
Tresp, V., Ahmad, S. and Neuneier, R. (1994). Training Neural Networks with Deficient Data. In: Cowan, J. D., Tesauro, G., and Alspector, J., eds., Advances in Neural
Information Processing Systems 6, San Mateo, CA, Morgan Kaufman.
Tresp, V., Hollatz J. and Ahmad, S. (1993). Network Structuring and Training Using
Rule-based Knowledge. In S. J. Hanson, J. D. Cowan and C. L. Giles, (Eds.), Advances
in Neural Information Processing Systems 5, San Mateo, CA: Morgan Kaufmann.
| 971 |@word version:1 advantageous:1 seems:1 retraining:2 tried:1 jacob:2 decomposition:4 covariance:2 tr:4 reduction:1 neuneier:2 current:1 nt:2 nowlan:3 location:1 toronto:1 miinchen:1 simpler:3 combine:1 introduce:2 manner:1 acquired:1 forgetting:1 expected:1 indeed:1 alspector:1 multi:1 brain:1 considering:2 becomes:2 confused:1 estimating:1 underlying:5 notation:1 kaufman:1 unified:5 ag:1 certainty:3 berkeley:1 act:1 classifier:6 unit:2 engineering:1 local:3 might:13 mateo:3 limited:1 bootstrap:1 digit:4 yio:2 cannot:1 close:2 selection:2 context:1 missing:7 center:2 zfe:1 independently:1 convex:1 simplicity:1 assigns:1 estimator:22 rule:8 gm:1 designing:1 pa:1 velocity:1 approximated:1 particularly:1 recognition:1 cesare:1 role:1 module:37 electrical:1 region:7 ahmad:6 valuable:1 yk:1 disease:2 environment:1 complexity:3 xio:2 personal:1 trained:10 depend:2 ithe:1 writer:4 basis:1 completely:1 easily:2 represented:2 train:2 stacked:1 forced:1 describe:1 modulators:1 sejnowski:1 neighborhood:1 modular:2 larger:1 plausible:1 say:1 otto:1 statistic:2 gi:3 g1:1 noisy:1 housing:1 propose:1 adaptation:2 combining:5 achieve:1 cluster:1 ring:1 depending:1 coupling:5 soc:1 ois:1 c:1 xly:1 indicate:1 direction:1 closely:1 filter:2 centered:1 generalization:2 decompose:1 preliminary:1 opt:1 biological:3 mathematically:1 extension:1 normal:1 exp:1 mapping:8 smallest:1 estimation:1 individually:1 weighted:2 mit:2 gaussian:4 modified:1 breiman:2 volker:6 structuring:1 derived:5 indicates:1 medication:1 am:1 dependent:4 nn:1 greatfully:1 i1:1 germany:1 special:1 integration:1 equal:3 identical:4 represents:1 unsupervised:1 few:3 randomly:1 individual:6 consisting:1 maintain:2 mlp:1 mixture:13 extreme:1 nl:3 re:2 haifa:1 uncertain:1 modeling:1 soft:1 giles:2 assignment:1 subset:1 technion:1 too:1 dependency:1 considerably:1 combined:2 density:9 international:1 pool:1 michael:1 quickly:1 concrete:1 squared:2 central:1 nm:1 thesis:1 expert:9 american:6 leading:1 li:3 account:1 de:1 explicitly:1 eyi:1 hollatz:2 extrapolation:2 capability:1 square:1 ni:17 variance:22 kaufmann:2 ensemble:1 correspond:1 ofthe:1 xli:4 handwritten:4 researcher:1 ed:5 pp:2 sejnowsky:1 obvious:1 associated:1 recall:1 knowledge:4 segmentation:1 reflected:1 improved:1 symptom:1 furthermore:3 correlation:1 hand:1 cohn:1 building:1 brown:1 unbiased:3 true:4 multiplier:1 consisted:1 assigned:2 i2:1 during:2 generalized:2 confusion:1 motion:3 novel:2 recently:1 superior:4 ji:1 discussed:1 hie:1 ai:1 llx:1 similarly:1 had:1 taniguchi:5 own:2 driven:1 tesauro:1 meta:1 seen:5 minimum:2 additional:4 somewhat:1 morgan:3 biasi:2 dashed:1 divided:2 calculates:1 prediction:5 regression:2 patient:1 expectation:1 vision:1 cmu:2 represent:4 sometimes:1 preserved:1 whereas:1 want:2 biased:1 deficient:1 cowan:3 seem:1 jordan:1 near:2 easy:1 switch:1 translates:1 motivated:1 hessian:1 action:2 useful:2 ylx:1 amount:2 locally:1 simplest:1 meir:2 estimated:2 tibshirani:2 vol:2 group:2 acknowledged:1 inverse:1 reasonable:1 appendix:1 layer:1 hi:5 summer:1 distinguish:1 nni:3 constraint:1 department:1 waibel:2 combination:2 perrone:3 poor:1 smaller:1 slightly:1 em:3 wi:1 biologically:1 intuitively:2 taken:2 equation:9 discus:1 know:3 flip:1 available:1 gaussians:1 enforce:1 bii:1 coin:1 hat:1 original:3 neuml:1 clustering:1 include:1 remaining:1 hahn:1 added:2 visiting:1 gradient:1 thank:1 seven:1 reason:1 modeled:1 relationship:3 providing:1 cij:2 unknown:3 hinton:1 communication:1 duff:1 competence:2 david:1 hanson:2 learned:1 pattern:3 ilxk:1 reliable:2 natural:1 residual:3 inversely:1 hm:1 coupled:1 tresp:13 faced:1 prior:2 acknowledgement:1 proportional:1 var:14 age:1 integrate:1 sufficient:1 classifying:1 uncorrelated:1 pi:1 gl:1 side:1 bias:3 distributed:1 calculated:1 evaluating:2 stand:3 author:1 adaptive:1 san:3 approximate:1 obtains:1 global:1 overfitting:3 assumed:1 gem:1 alternatively:2 don:1 continuous:6 un:1 pretty:1 reality:1 table:2 robust:1 ca:3 correlated:3 improving:1 excellent:3 complex:4 european:6 main:2 linearly:1 motivation:1 noise:2 n2:6 lc:1 ylo:3 xl:1 candidate:2 ib:1 weighting:38 learns:1 ix:1 dozen:1 bad:1 gating:3 phd:1 boston:1 ilx:3 eij:2 likely:1 simply:1 visual:1 lagrange:1 xlo:3 partially:1 gender:1 corresponds:3 conditional:1 goal:2 included:1 determined:1 except:1 averaging:7 hampshire:2 catastrophic:1 siemens:2 perceptrons:1 select:2 formally:2 latter:2 incorporate:1 dept:2 tested:1 ex:1 |
7,212 | 972 | A model of the hippocampus combining selforganization and associative memory function.
Michael E. Hasselmo, Eric Schnell
Joshua Berke and Edi Barkai
Dept. of Psychology, Harvard University
33 Kirkland St., Cambridge, MA 02138
hasselmo@katla.harvard.edu
Abstract
A model of the hippocampus is presented which forms rapid self-organized representations of input arriving via the perforant path, performs
recall of previous associations in region CA3, and performs comparison
of this recall with afferent input in region CA 1. This comparison drives
feedback regulation of cholinergic modulation to set appropriate
dynamics for learning of new representations in region CA3 and CA 1.
The network responds to novel patterns with increased cholinergic modulation, allowing storage of new self-organized representations, but
responds to familiar patterns with a decrease in acetylcholine, allowing
recall based on previous representations. This requires selectivity of the
cholinergic suppression of synaptic transmission in stratum radiatum of
regions CA3 and CAl, which has been demonstrated experimentally.
1
INTRODUCTION
A number of models of hippocampal function have been developed (Burgess et aI., 1994;
Myers and Gluck, 1994; Touretzky et al., 1994), but remarkably few simulations have
addressed hippocampal function within the constraints provided by physiological and anatomical data. Theories of the function of specific subregions of the hippocampal formation often do not address physiological mechanisms for changing dynamics between
learning of novel stimuli and recall of familiar stimuli. For example, the afferent input to
the hippocampus has been proposed to form orthogonal representations of entorhinal
activity (Marr, 1971; McNaughton and Morris, 1987; Eichenbaum and Buckingham,
1990), but simulations have not addressed the problem of when these representations
78
Michael E. Hasselmo. Eric Schnell. Joshua Berke. Edi Barkai
should remain stable, and when they should be altered. In addition, models of autoassociative memory function in region CA3 (Marr, 1971; McNaughton and Morris, 1987;
Levy, 1989; Eichenbaum and Buckingham, 1990) and heteroassociative memory function
at the Schaffer collaterals projecting from region CA3 to CAl (Levy, 1989; McNaughton,
1991) require very different activation dynamics during learning versus recall.
Acetylcholine may set appropriate dynamics for storing new information in the cortex
(Hasselmo et aI., 1992, 1993; Hasselmo, 1993, 1994; Hasselmo and Bower, 1993). Acetylcholine has been shown to selectively suppress synaptic transmission at intrinsic but
not afferent fiber synapses (Hasselmo and Bower, 1992), to suppress the neuronal adaptation of cortical pyramidal cells (Hasselmo et aI., 1994; Barkai and Hasselmo, 1994), and
to enhance long-term potentiation of synaptic potentials (Hasselmo, 1994b). Models
show that suppression of synaptic transmission during learning prevents recall of previously stored information from interfering with the storage of new information (Hasselmo
et al., 1992, 1993; Hasselmo, 1993, 1994a), while cholinergic enhancement of synaptic
modification enhances the rate of learning (Hasselmo, 1994b).
Feedback regulation of cholinergic modulation may set the appropriate level of cholinergic modulation dependent upon the novelty or familiarity of a particular input pattern.
We have explored possible mechanisms for the feedback regulation of cholinergic modulation in simulations of region CAl (Hasselmo and Schnell, 1994) and region CA3. Here
we show that self-regulated learning and recall of self-organized representations can be
obtained in a network simulation of the hippocampal formation. This model utilizes selective cholinergic suppression of synaptic transmission in stratum radiatum of region CA3,
which has been demonstrated in brain slice preparations of the hippocampus.
2
METHODS
2.1. SIMPLIFIED REPRESENTA nON OF HIPPOCAMPAL NEURONS.
In place of the sigmoid input-output functions used in many models, this model uses a
simple representation in which the output of a neuron is not explicitly constrained, but the
total network activity is regulated by feedback from inhibitory interneurons and adaptation due to intracellular calcium concentration. Separate variables represent pyramidal
cell membrane potential a, intracellular calcium concentration c, and the membrane potential of inhibitory interneurons h:
l1a i = Ai -l1 ai - J..l.C +
L Wijg(aj - e) - Hikg(h k - e h)
j
I1c?I = 'Vg(a
.i?
I
I1hk
e )- Qc
C
= IWkjg(aj-eo)-l1hk- IHk/g(h/-e)
j
where A = afferent input, "
/
=passive decay of membrane potential, Il =strength of cal-
A Model of Hippocampus
79
cium-dependent potassium current (proportional to intracellular calcium), Wij = excitatory
recurrent synapses (longitudinal association path tenninating in stratum radiatum), gO is a
threshold linear function proportional to the amount by which membrane potential
exceeds an output threshold 00 or threshold for calcium current Oc' 'Y = strength of voltagedependent calcium current, n = diffusion constant of calcium, Wki = excitatory synapses
inhibitory interneurons, Hilc = inhibitory synapses from interneurons to pyramidal cells,
Hk}= inhibitory synapses between interneurons. This representation gives neurons adaptation characteristics similar to those observed with intracellular recording (Barkai and Hasselmo, 1994), including a prominent afterhyperpolarization potential (see Figure 1).
An
B
.... ..
~
....J
'N~JJL
C
lO
\..--14
Figure 1. Comparison of pyramidal cell model with experimental data.
In Figure I, A shows the membrane potential of a modeled pyramidal cell in response to
simulated current injection. Output of this model is a continuous variable proportional to
how much membrane potential exceeds threshold. This is analogous to the reciprocal of
interspike interval in real neuronal recordings. Note that the model displays adaptation
during current injection and afterhyperpolarization afterwards, due to the calcium-dependent potassium current. B shows the intracellularly recorded membrane potential in a pirifonn cortex pyramidal cell, demonstrating adaptation of firing frequency due to
activation of calcium-dependent potassium current. The firing rate falls off in a manner
similar to the smooth decrease in firing rate in the simplified representation. C shows an
intracellular recording illustrating long-tenn afterhyperpolarization caused by calcium
influx induced by spiking of the neuron during current injection.
2.2. NETWORK CONNECTIVITY
A schematic representation of the network simulation of the hippocampal fonnation is
shown in Figure 2. The anatomy of the hippocampal fonnation is summarized on the left
in A, and the function of these different subregions in the model is shown on the right in
B. Each of the subregions in the model contained a population of excitatory neurons with
a single inhibitory interneuron mediating feedback inhibition and keeping excitatory
activity bounded. Thus, the local activation dynamics in each region follow the equations
presented above. The connectivity of the network is further summarized in Figure 3 in the
Results section. A learning rule of the Hebbian type was utilized at all synaptic connections, with the exception of the mossy fibers from the dentate gyrus to region CA3, and the
connections to and from the medial septum. Self-organization of perforant path synapses
was obtained through decay of synapses with only pre or post-synaptic activity, and
growth of synapses with combined activity. Associative memory function at synapses
80
Michael E. Hasse/mo, Eric Schnell, Joshua Berke, Edi Barkai
arising from region CA3 was obtained through synaptic modification during cholinergic
suppression of synaptic transmission.
B
Entorhinal cortex
""""""""""
??
?
???
: Self-organized
: representation
: r-----..L.\
?
~
Comparison
.L-______~~.-----~~~--~~
Feedback regulation of
cholinergic modulation
Regulation of
learning dynamics
Figure 2. Schematic representation of hippocampal circuitry
and the corresponding function of connections in the model.
2.3. CHOLINERGIC MODULAnON
The total output from region CAl determined the level of cholinergic modulation within
both region CA3 and CAl, with increased output causing decreased modulation. This is
consistent with experimental evidence suggesting that activity in region CAl and region
CA3 can inhibit activity in the medial septum, and thereby downregulate cholinergic modulation. This effect was obtained in the model by excitatory connections from region CAl
to an inhibitory interneuron in the medial septum, which suppressed the activity of a cholinergic neuron providing modulation to the full network. When levels of cholinergic
modulation were high, there was strong suppression of synaptic transmission at the excitatory recurrent synapses in CA3 and the Schaffer collaterals projecting from region CA3 to
CAL This prevented the spread of activity due to previous learning from interfering with
self-organization. When levels of cholinergic modulation were decreased, the strength of
synaptic transmission was increased, allowing associative recall to dominate. Cholinergic
modulation also increased the rate of synaptic modification and depolarized neurons.
2.4. TESTS OF SELF-REGULATED LEARNING AND RECALL
Simulations of the full hippocampal network evaluated the response to the sequential presentation of a series of highly overlapping activity patterns in the entorhinal cortex. Recall
was tested with interspersed presentation of degraded versions of previously presented
activity patterns. For effective recall, the pattern of activity in entorhinal cortex layer IV
evoked by degraded patterns matched the pattern evoked by the full learned version of
these patterns. The function of the full network is illustrated in Figure 3. In simulations
A Model of Hippocampus
81
focused on region CA3, activity patterns were induced sequentially in region CA3, representing afferent input from the entorhinal cortex. Different levels of external activation of
the cholinergic neuron resulted in different levels of learning of new overlapping patterns.
These results are illustrated in Figure 4.
2.5. BRAIN SLICE EXPERIMENTS
The effects in the simulations of region CA3 depended upon the cholinergic suppression
of synaptic transmission in stratum radiatum of this region The cholinergic suppression of
glutamatergic synaptic transmission in region CA3 was tested in brain slice preparations
by analysis of the influence of the cholinergic agonist carbachol on the size of field potentials elicited by stimulation of stratum radiatum. These experiments used techniques similar to previously published work in region CAl (Hasselmo and Schnell, 1994).
3 RESULTS
In the full hippocampal simulation, input of an unfamiliar pattern to entorhinal cortex
layer II resulted in high levels of acetylcholine. This allowed rapid self-organization of
the perforant path input to the dentate gyrus and region CAl. Cholinergic suppression of
synaptic transmission in region CAl prevented recall from interfering with self-organization. Instead, recurrent collaterals in region CA3 stored an autoassociative representation
of the input from the dentate gyrus to region CA3, and connections from CA3 to CA 1
stored associations between the pattern of activity in CA3 and the associated self-organized representation in region CAl.
identity
" self-org "matrix ~
,auto-"
,,~>
M assoc
u .....
?
?
'at)
u
u
c >.
--I.~ ?
c
?
c
?
:.a.?:i
Self-org ~
iden~ity
.9
hetero.9 hetero8
matrix
~
assoc
~ assoc
C!
~
~
?
~
111111 I "T I' , If
,r
2 I I I II
n
I I
ld II
r I" I 'I"'"'I II , , ,
Q)2dl' r II' I I ' l l I
~ 311111 I' I I I n
"
j
4
3d
4d
ld
2d
n,n
II
II
II
'f
I I II I n
I
II I
I
,
,
II
,
I
1I
'I
I
'I
.. ,
II f'l't?
U i
I I Itt
~
I.Ll
r"
U
I
"
H
,
I
I
U
I I 1 II r
~'
,,
1
,
'(
( U.
"
I I
I
1'1 I I
I I I n
I"'" , 'I II I II
1 I ' ,I
I
l 11 I J1
't III I'
?"
'I . .
I
I
, I I II
I II ? .1
l " I II
I III I I
W.
Jl 1
lU
Neuron #
Figure 3. Activity in each subregion of the full network simulation of the hippocampal
formation during presentation of a sequence of activity patterns in entorhinal cortex.
82
Michael E. Hasselmo, Eric Schnell, Joshua Berke, Edi Barkai
In Figure 3. width of the lines represents the activity of each neuron at a particular time
step. As seen here. the network forms a self-organized representation of each new pattern
consisting of active neurons in the dentate gyrus and region CAL At the same time. an
association is formed between the self-organized representation in region CAl and the
same afferent input pattern presented to entorhinal cortex layer IV. Four overlapping patterns (1-4) are presented sequentially. each of which results in learning of a separate selforganized representation in the dentate gyrus and region CAl. with an association formed
between this representation and the full input pattern in entorhinal cortex.
The recall characteristics of the network are apparent when degraded versions of the afferent input patterns are presented in the sequence (ld-4d). This degraded afferent input
weakly activates the same representations previously formed in the dentate gyrus. Recurrent excitation in region CA3 enhances this activity. giving robust recall of the full version
of this pattern. This activity then reaches CA 1. where it causes strong activation if it
matches the pattern of afferent input from the entorhinal cortex. Strong activation in
region CAl decreases cholinergic modulation. preventing formation of a new representation and allowing recall to dominate. Strong activation of the representation stored in
region CAl then activates the full representation of the pattern in entorhinal cortex layer
IV. Thus. the network can accurately recall each of many highly overlapping patterns.
The effect of cholinergic modulation on the level of learning or recall can be seen more
clearly in a simulation of auto-associative memory function in region CA3 as shown in
Figure 4. Each box shows the response of the network to sequential presentation of full
and degraded versions of two highly overlapping input patterns. The width of the black
traces represents the activity of each of 10 CA3 pyramidal cells during each simulation
step. In the top row. level of cholinergic modulation (ACh) is plotted. In A. external activation of the cholinergic neuron is absent. so there is no cholinergic suppression of synaptic transmission. In this case. the first pattern is learned and recalled properly. but
subsequent presentation of a second overlapping pattern results only in recall of the previously learned pattern. In B. with greater cholinergic suppression. recall is suppressed sufficiently to allow learning of a combination of the two input patterns. Finally. in C. strong
cholinergic suppression prevents recall. allowing learning of the new overlapping pattern
to dominate over the previously stored pattern.
A
Stored
patterns
???
??
??
?
?
-.gN.g
N
-
ACh
Inhib
Q\
.- .--.
ACh input = 0.0
..... ......
,11",,11.
.....
B
ACh input
=0.15
C
__
ACh input
=0.3
...11'_..,.......,.. ......
111 ... 111...... 11 . . . ."'... 11 ? ?111 ??,
'
I ? ? " '? ? _ _
'111"
.';::
~
~
?'. _"11.. ? .'.? .
.,~
. ",.
.u. . . . .... ,
0
. . . . . . . . . . 111. . . .... .
N
Figure 4. Increased cholinergic suppression of synaptic transmission in region CA3
causes greater learning of new aspects of afferent input patterns.
A Model of Hippocampus
83
Extracellular recording in brain slice preparations of hippocampal region CA3 have demonstrated that perfusion of the cholinergic agonist carbachol strongly suppresses synaptic
potentials recorded in stratum radiatum, as shown in Figure 5. In contrast, suppression of
synaptic transmission at the afferent fiber synapses arising from entorhinal cortex is much
weaker. At a concentration of 20J..tM, carbachol suppressed synaptic potentials in stratum
radiatum on average by 54.4% (n=5). Synaptic potentials elicited in stratum lacunosum
were more weakly suppressed, with an average suppression of28%.
Control
Carbachol
(20JlM)
Wash
Figure 5. Cholinergic suppression of synaptic transmission in stratum radiatum of CA3.
4
DISCUSSION
In this model of the hippocampus, self-organization at perforant path synapses forms compressed representations of specific patterns of cortical activity associated with events in
the environment. Feedback regulation of cholinergic modulation sets appropriate dynamics for learning in response to novel stimuli, allowing predominance of self-organization,
and appropriate dynamics for recall in response to familiar stimuli, allowing predominance of associative memory function. This combination of self-organization and associative memory function may also occur in neocortical structures. The selective cholinergic
suppression of feedback and intrinsic synapses has been proposed to allow self-organization of feedforward synapses while feedback synapses mediate storage of associations
between higher level representations and activity in primary cortical areas (Hasselmo,
1994b). This previous proposal could provide a physiological justification for a similar
mechanism utilized in recent models (Dayan et al., 1995). Detailed modeling of cholinergic effects in the hippocampus provides a theoretical framework for linking the considerable behavioral evidence for a role of acetylcholine in memory function (Hagan and
Morris, 1989) to the neurophysiological evidence for the effects of acetylcholine within
cortical structures (Hasselmo and Bower, 1992; 1993; Hasselmo, 1994a, 1994b).
Acknowledgements
This work supported by a pilot grant from the Massachusetts Alzheimer's Disease
Research Center and by an NIMH FIRST award MH52732-01.
References
Barkai E, Hasselmo ME (1994) Modulation of the input/output function of rat piriform
cortex pyramidal cells. J. Neurophysiol. 72: 644-658.
84
Michael E. Hasselmo, Eric Schnell, Joshua Berke, Edi Barkai
Barkai E, Bergman RE, Horwitz G, Hasselmo ME (1994) Modulation of associative memory function in a biophysical simulation of rat pirifonn cortex. J. Neurophysiol. 72:659677.
Burgess N, Reece M, O'Keefe J (1994) A model of hippocampal function. Neural Networks 7: 1065-1081.
Dayan P, Hinton GE, Neal RM and Zemel RS (1995) The Helmholtz machine. Neural
computation in press.
Eichenbaum, H. and Buckingham, J. (1990) Studies on hippocampal processing: experiment, theory and model. In: Learning and computational neuroscience: foundations of
adaptive networks, M. Gabriel and J. Moore, eds., Cambridge, MA: MIT Press.
Hagan, JJ and Morris, RGM (1989) The cholinergic hypothesis of memory: A review of
animal experiments. In Psychopharmacology of the Aging Nervous System, L.L. Iversen,
S.D. Iversen and S.H. Snyder, eds. New York: Plenum Press, p. 237-324.
Hasselmo, M.E. (1993) Acetylcholine and learning in a cortical associative memory. Neural Compo 5: 22-34.
Hasselmo ME (1994a) Runaway synaptic modification in models of cortex: Implications
for Alzheimer's disease. Neural Networks 7: 13-40.
Hasselmo ME (1994b) Neuromodulation and cortical function. Behav. Brain Res. in press
Hasselmo ME, Anderson, BP and Bower, JM (1992) Cholinergic modulation of cortical
associative memory function. J. Neurophysiol. 67(5): 1230-1246.
Hasselmo ME, Bower JM (1992) Cholinergic suppression specific to intrinsic not afferent
fiber synapses in rat pirifonn (olfactory) cortex. J. Neurophysiol. 67(5): 1222-1229.
Hasselmo ME, Bower JM (1993) Acetylcholine and memory. Trends Neurosci 16:218222.
Hasselmo ME, Barkai E, Horwitz G, Bergman RE (1993) Modulation of neuronal adaptation and cortical associative memory function. In: Computation and Neural Systems II
(Eeckman F, Bower JM, ed). Norwell, MA: Kluwer Academic Publishers.
Hasselmo ME, Schnell E (1994) Laminar selectivity of the cholinergic suppression of synaptic transmission in rat hippocampal region CAl: Computational modeling and brain
slice physiology. J. Neurosci. 14: 3898-3914.
Levy WB (1989) A computational approach to hippocampal function. In: Computational
models of learning in simple neural systems (Hawkins RD, Bower GH, ed), pp. 243-305.
Orlando, FL: Academic Press.
Myers CE and Gluck M (1994) Context, conditioning and hippocampal rerepresentation
in animal learning. Behav. Neurosci. 108: 835-847.
Marr 0 (1971) Simple memory: A theory for archicortex. Phil. Trans. Roy. Soc. B
B262:23-81
McNaughton BL (1991) Associative pattern completion in hippocampal circuits: New
evidence and new questions. Brain Res. Rev. 16:193-220.
McNaughton BL, Morris RGM (1987) Hippocampal synaptic enhancement and infonnation storage within a distributed memory system. Trends Neurosci. 10:408-415.
Touretzky OS, Wan HS and Redish AD (1994) Neural representation of space in rats and
robots. In Zurada JM and Marks RJ (eds) Computational Intelligence: Imitating life.
IEEE Press.
| 972 |@word h:1 illustrating:1 version:5 selforganization:1 hippocampus:9 simulation:13 r:1 heteroassociative:1 thereby:1 ld:3 series:1 zurada:1 longitudinal:1 current:8 activation:8 buckingham:3 subsequent:1 j1:1 interspike:1 medial:3 tenn:1 intelligence:1 nervous:1 reciprocal:1 compo:1 provides:1 rgm:2 org:2 behavioral:1 olfactory:1 manner:1 rapid:2 brain:7 jm:5 psychopharmacology:1 provided:1 bounded:1 wki:1 matched:1 circuit:1 suppresses:1 developed:1 jlm:1 growth:1 assoc:3 rm:1 control:1 grant:1 local:1 depended:1 aging:1 path:5 modulation:21 firing:3 black:1 evoked:2 area:1 physiology:1 pre:1 cal:19 storage:4 context:1 influence:1 demonstrated:3 center:1 phil:1 go:1 focused:1 qc:1 rule:1 marr:3 dominate:3 mossy:1 population:1 ity:1 mcnaughton:5 analogous:1 justification:1 plenum:1 us:1 hypothesis:1 bergman:2 harvard:2 trend:2 helmholtz:1 roy:1 utilized:2 intracellularly:1 hagan:2 observed:1 role:1 region:39 decrease:3 inhibit:1 disease:2 environment:1 nimh:1 schnell:8 dynamic:8 weakly:2 upon:2 eric:5 neurophysiol:4 fiber:4 reece:1 effective:1 zemel:1 formation:4 apparent:1 compressed:1 associative:11 sequence:2 myers:2 biophysical:1 adaptation:6 causing:1 combining:1 potassium:3 enhancement:2 transmission:15 ach:5 perfusion:1 recurrent:4 completion:1 subregion:1 soc:1 strong:5 anatomy:1 hetero:1 runaway:1 require:1 potentiation:1 orlando:1 sufficiently:1 hawkins:1 dentate:6 mo:1 circuitry:1 infonnation:1 predominance:2 hasselmo:31 mit:1 clearly:1 activates:2 representa:1 acetylcholine:8 properly:1 hk:1 contrast:1 suppression:18 ihk:1 dependent:4 dayan:2 selective:2 wij:1 animal:2 constrained:1 jjl:1 field:1 represents:2 stimulus:4 few:1 resulted:2 iden:1 familiar:3 consisting:1 organization:8 interneurons:5 wijg:1 highly:3 cholinergic:39 implication:1 norwell:1 collateral:3 orthogonal:1 iv:3 re:4 i1c:1 plotted:1 theoretical:1 increased:5 modeling:2 wb:1 gn:1 ca3:27 stored:6 combined:1 st:1 stratum:9 pirifonn:3 off:1 michael:5 enhance:1 connectivity:2 recorded:2 wan:1 external:2 suggesting:1 potential:13 summarized:2 afterhyperpolarization:3 redish:1 explicitly:1 afferent:12 caused:1 ad:1 elicited:2 il:1 formed:3 degraded:5 characteristic:2 accurately:1 agonist:2 lu:1 drive:1 published:1 horwitz:2 synapsis:16 reach:1 touretzky:2 synaptic:26 ed:5 frequency:1 pp:1 associated:2 pilot:1 massachusetts:1 recall:21 organized:7 higher:1 follow:1 response:5 evaluated:1 box:1 strongly:1 anderson:1 o:1 overlapping:7 berke:5 aj:2 barkai:10 effect:5 perforant:4 hasse:1 moore:1 neal:1 illustrated:2 ll:1 during:7 self:19 width:2 excitation:1 oc:1 rat:5 hippocampal:19 prominent:1 neocortical:1 performs:2 l1:1 gh:1 passive:1 novel:3 archicortex:1 sigmoid:1 stimulation:1 spiking:1 conditioning:1 inhib:1 interspersed:1 association:6 jl:1 linking:1 kluwer:1 unfamiliar:1 cambridge:2 ai:5 eeckman:1 rd:1 stable:1 robot:1 cortex:17 inhibition:1 recent:1 selectivity:2 life:1 joshua:5 seen:2 greater:2 eo:1 novelty:1 ii:19 afterwards:1 full:10 fonnation:2 rj:1 hebbian:1 exceeds:2 smooth:1 match:1 academic:2 long:2 post:1 prevented:2 award:1 schematic:2 represent:1 cell:8 proposal:1 addition:1 remarkably:1 addressed:2 interval:1 decreased:2 pyramidal:8 publisher:1 depolarized:1 recording:4 induced:2 alzheimer:2 feedforward:1 iii:2 psychology:1 burgess:2 kirkland:1 tm:1 absent:1 septum:3 york:1 cause:2 jj:1 behav:2 autoassociative:2 gabriel:1 detailed:1 amount:1 subregions:3 morris:5 gyrus:6 inhibitory:7 neuroscience:1 arising:2 anatomical:1 snyder:1 four:1 threshold:4 demonstrating:1 changing:1 ce:1 diffusion:1 place:1 utilizes:1 layer:4 fl:1 display:1 laminar:1 activity:22 strength:3 occur:1 constraint:1 bp:1 influx:1 aspect:1 injection:3 eichenbaum:3 extracellular:1 combination:2 membrane:7 remain:1 suppressed:4 voltagedependent:1 modification:4 rev:1 projecting:2 imitating:1 equation:1 previously:6 mechanism:3 neuromodulation:1 ge:1 appropriate:5 top:1 iversen:2 giving:1 bl:2 question:1 concentration:3 primary:1 responds:2 enhances:2 regulated:3 separate:2 simulated:1 me:9 modeled:1 providing:1 piriform:1 regulation:6 mediating:1 trace:1 suppress:2 calcium:9 allowing:7 neuron:12 hinton:1 schaffer:2 edi:5 connection:5 recalled:1 learned:3 trans:1 address:1 pattern:34 including:1 memory:16 event:1 representing:1 altered:1 auto:2 review:1 acknowledgement:1 radiatum:8 proportional:3 versus:1 vg:1 foundation:1 consistent:1 storing:1 interfering:3 lo:1 row:1 excitatory:6 supported:1 keeping:1 arriving:1 allow:2 weaker:1 fall:1 distributed:1 slice:5 feedback:9 cortical:8 preventing:1 adaptive:1 simplified:2 sequentially:2 active:1 continuous:1 itt:1 robust:1 ca:4 selforganized:1 carbachol:4 spread:1 intracellular:5 neurosci:4 mediate:1 allowed:1 neuronal:3 bower:8 levy:3 familiarity:1 specific:3 explored:1 decay:2 physiological:3 evidence:4 dl:1 intrinsic:3 sequential:2 keefe:1 entorhinal:12 wash:1 interneuron:2 gluck:2 neurophysiological:1 prevents:2 contained:1 ma:3 identity:1 presentation:5 considerable:1 experimentally:1 determined:1 total:2 experimental:2 exception:1 selectively:1 mark:1 preparation:3 dept:1 tested:2 |
7,213 | 973 | Catastrophic Interference in Human
Motor Learning
Tom Brashers-Krug, Reza Shadmehr t , and Emanuel Todorov
Dept. of Brain and Cognitive Sciences, M. I. T., Cambridge, MA 02139
tCurrently at Dept. of Biomedical Eng., Johns Hopkins Univ., Baltimore, MD 21205
Email: tbk@ai.mit.edu, reza@bme.jhu.edu, emo@aLmit.edu
Abstract
Biological sensorimotor systems are not static maps that transform
input (sensory information) into output (motor behavior). Evidence from many lines of research suggests that their representations are plastic, experience-dependent entities. While this plasticity is essential for flexible behavior, it presents the nervous system
with difficult organizational challenges. If the sensorimotor system
adapts itself to perform well under one set of circumstances, will it
then perform poorly when placed in an environment with different
demands (negative transfer)? Will a later experience-dependent
change undo the benefits of previous learning (catastrophic interference)? We explore the first question in a separate paper in this
volume (Shadmehr et al. 1995). Here we present psychophysical
and computational results that explore the question of catastrophic
interference in the context of a dynamic motor learning task. Under some conditions, subjects show evidence of catastrophic interference. Under other conditions, however, subjects appear to be
immune to its effects. These results suggest that motor learning
can undergo a process of consolidation. Modular neural networks
are well suited for the demands of learning multiple input/output
mappings. By incorporating the notion of fast- and slow-changing
connections into a modular architecture, we were able to account
for the psychophysical results.
20
1
Tom Brashers-Krug, Reza Shadmelzr, Emanuel Todorov
Introduction
Interacting physically with the world changes the dynamics of one's limbs. For
example, when holding a heavy load, a different pattern of muscular activity is
needed to move one's arm along a particular path than when not holding a load.
Previous work in our laboratory has shown that humans learn a novel dynamic
task by forming an internal model of the new inverse dynamics of thier limbs.
(Shadmehr and Mussa-Ivaldi 1994, Shadmehr et aI, 1995). Preliminary evidence
suggests that subjects can retain one of these internal models over time (BrashersKrug, et al. 1994). Humans are required, however, to move their limbs effectively
under a large number of dynamic conditions. Are people able to learn and store
an inverse dynamic model appropriate for each condition, or do they form such
models from scratch as they need them? In particular, can learning a new inverse
dynamic model overwrite or displace a previous model? We will present evidence
that certain conditions must be met before a subject is able to retain more than
one inverse dynamic model in a given experimental context. These conditions can
be modeled as leading to a process of consolidation, whereby learning is transfered
from vulnerable, low-capacity storage to a long-term, high-capacity storage.
2
Experimental Protocol
We have developed a motor learning paradigm that allows us to alter the dynamics
of a subject's arm and so to monitor a subject's ability to learn dynamic tasks.
A subject moves the handle on the free end of a two-link planar robot arm-called
a manipulandum-to guide a cursor to a series of targets displayed on a computer
screen (fig la). The position and velocity of the handle of the manipulandum are
recorded at ten-millisecond intervals and are used to deliver state-dependent forces
to the subject's hand. In order to test a subject's ability to learn a novel dynamic
task, we present the subject with a viscous force field as slhe moves from one
target to the next (fig Ib). Initially, such forces perturb the subject's movements,
causing them to depart from the smooth, straight-line trajectories of the baseline
condition (i .e., the condition before the viscous forces were presented) (figs lc,ld).
The extent of learning is measured as the degree to which a subject's movements
in the force field over time come to resemble that subject's baseline movements.
We have shown in previous work that subjects adapt to the imposed force fileds by
forming a predictive model of the velocity-dependent forces, and that subjects use
this inverse dynamic model to control their arms in what appears to be a feedforward
manner (Shadmehr and Mussa-Ivaldi 1994).
3
3.1
Psychophysical Findings
Catastrophic Interference
Here, we employed this paradigm to explore the consequences of learning two different dynamic tasks in a row. In an initial series of experiments, we allowed twelve
subjects to learn to move the manipulandum in a first force field (Field A) for approximately 5 minutes. Immediately after this first set of movements, we presented
the subjects with an anti-correlated force field (Field B). For example, if we pre-
Catastrophic Interference in Human Motor Learning
<Y'/'/ .
C
21
\
D
Figure 1: A: The experimental setup. B: An example of a viscous field, plotted in velocity
space (mm/sec). The arrows indicate the direction and magnitude of the forces exerted
by the manipulandum on the subject's hand at each location in velocity space. C: One
subject's trajectories before forces were introduced. Targets are indicated by the open
circles. D: Trajectories immediately after the force field in (B) was presented.
sen ted the counter-clockwise curl field depicted in fig. 1B as Field A, we would next
present a clockwise curl field as Field B. Half the subjects learned the clockwise curl
field first and the counter-clockwise field second; the other halflearned the two fields
in the reverse order. (The first field will be referred to as Field A and the second
field as Field B, whichever field was learned first.) The subjects' mean performance
in Field B was worse (p< 0.0001, paired t-test) than in Field A. This phenomenon
has been called negative transfer in the psychophysical literature. Negative transfer
in this motor learning paradigm is explored more fully in a separate paper in this
volume (Shadmehr et aI, 1995). In that paper, we suggested that this negative
transfer could result from the fact that the same neural elements are learning both
tasks. We predicted that, if this is the case, learning to move in Field B would
interfere with a subject's ability to retain an inverse dynamic model of Field A.
Learning to move in Field B would, in effect, cause subjects to "unlearn" Field A,
resulting in catastrophic interference.
In order to test this prediction, we compared the improvement in performance from
one day to the next of two groups of subjects, with twelve subjects in each group.
The subjects in the control group learned to move in one force field on Day One
and were then tested on Day Two in the same force field. The subjects in the
experimental group learned two separate force fields in a row on Day One and were
then tested on Day Two in the first force field they learned. The experimental
group retained significantly less of their learning (p< 0.01, paired t-test) from Day
One to Day Two than the control group (figs 2a,2b) . In other words, learning the
second force field resulted in catastrophic interference. (The question of whether
this represents a storage or a retrieval phenomenon is beyond the scope of this
paper.)
22
3.2
Tom Brashers-Krug, Reza Shadmehr, Emanuel Todorov
Consolidation
Having found evidence for catastrophic interference, we wanted to know whether
there were circumstances under which dynamic motor learning was immune to being
functionally erased by subsequent learning. We therefore tested two further groups
of subjects. We allowed these subjects to practice longer in one field before they
learned the second field. We also allowed 24 hours to pass between the time subjects
first learned one field and when they learned the second field . The subjects in
the experimental group (n = 10) practiced in one force field for approximately 15
minutes on Day One. They returned on Day Two and practiced in the same force
field for five more minutes. They were then allowed to practice in a second force
field for 15 minutes. By the end of this fifteen minutes, they were performing in the
second field at a level comparable to the level they acheived in the first force field.
We had the subjects return on Day Three, when they were tested for their retention
of the first field they learned. We compared the difference in performance on Day
Two and Day Three of this experimental group with that of a control group (n =
9) who followed the same protocol in all respects except that they never learned a
second force field . In this way we could determine whether learning the second field
resulted in a decrement in performance for the experimental group when compared
with the control group.
Under these conditions, we found no difference in the retention of learning between
the experimental and control groups (fig 2c, 2d) . That is, learning the second
field under these conditions no longer resulted in catastrophic interference. What
subjects had learned about the first field had become resistant to such interference.
It had become consolidated.
We can not tell from these experiments whether such consolidation is the result of
the increased practice in the first field, or whether it is the result of the 24 hours
that elapsed between when the first field was first learned and when the second field
was learned. There is evidence that increased practice in a motor task can engage
different neural circuits than those that are active during initial learning (Jenkins,
et al 1994). The shift to "practiced" circuits may represent the neural substrate of
consolidation. There is also evidence that time can be an important factor in the
consolidation of skill learning. (Karni and Sagi 1993) In the next section, we present
a model of our results that assumes that time is the key variable in the consolidation
of motor learning. The model could also be applied to a practice-based model of
consolidation with minor modifications.
4
Computational Modeling of the Experimental Results
In order to model the results presented above we need a network that learns to
compute an appropriate control signal Y given the current state and the desired
next state X of the plant. More precisely, it needs to compute a mapping from joint
angles 0, joint velocities iJ, and desired joint accelerations jj to torques. Several approaches for solving this problem have been proposed. One way to learn a mapping
from X into Y is to use direct inverse modeling: apply a control signal, measure the
next state of the plant, and use the current state, new state, and control signal as
a training pair for the controller. This approach is not suitable for explaining non-
23
Catastrophic Interference ill Human Motor Learning
0.95
0.95
0.85
0.75
~
","
0.75
A
0.65
0
50
0.85
100
0
150
0.95
0.85
0.85
0.75
0
50
100
150
B
0.65
0.95
C
~
50
100
150
~
D
0.75
0
50
100
150
Figure 2: Plots of average learning curves. The correlation of trajectories in the force
field to baseline trajectories (before the force field was applied) is plotted as a function
of movement number. A: Learning curves for the first experimental group. Dark curve:
learning curve on Day One in Field A . Light curve: learning curve in Field A on Day Two.
This group learned Field B immediately after Field A on Day One (learning curve for
Field B not shown). Note minimal improvement from Day One to Day Two. B: Learning
curves for the first control group. Dark curve: learning curve in Field A on Day One.
Light curve: learning curve in Field A on Day Two. Control group never learned Field
B. Note significant improvement from Day One to Day Two. C: Second experimental
group. Dark line: learning curve in Field A on Day Two immediately before learning
Field B (Field B curve not shown). Light line: learning curve in Field A on Day Three.
D: Second control group. Dark and light lines: learning curves in FIeld A on Day Two
and Three, respectively. Note the similarity of the curves in C and D. This indicates that
learning Field B did not significantly affect the experimental group's retention of Field A.
All curves in C and D are significantly higher than curves for the initial learning of Field
A on Day One.
convex mappings, however. The learning situation we must model is non-convex:
we change the dynamic environment the controller operates in by presenting force
fileds, and so there will be different Y values corresponding to any X value. A different approach that solves the non convexity problem is distal supervised learning
(Jordan and Rumelhart 1992): produce control signals, observe the new state of
the plant, and use that information to train a forward model that maps actions into
states; then learn a controller, using the forward model to backpropagate error from
observable cartesian coordinates to the unknown control space. Distal supervised
learning solves the non convexity problem by learning one correct value of Y for each
X. But that can not explain the consolidation of learning - when the force field
changes back to something already learned, our controller should rapidly recover its
performance in that old field, meaning that it should retain information about all
Y s that a particular X can map into.
An architecture that seems to have most of the desirable properties discussed above
is the Mixture of Experts (ME) model (Jordan and Jacobs 1994): several experts
learn a mapping from X to Y . A separate gating network selects an expert which is
most likely to be correct at each moment. Such a model has been used previously
(Jacobs and Jordan 1993) to learn a feedforward controller for a 2-joint planar arm
operating under different loads. In their model however they assumed that the
24
Tom Brashers-Krug, Reza Shadmelzr, Emanuel Todorov
identity of the load is known to the controller. The subjects in our study were
not given any explicit cues about the identity of the fields they were learning. The
mixture of experts cannot be used directly here because it decides which expert
to select based on a soft partitioning of the input space, and in our experiment
any force field is active over any portion of the input space at different moments
in time. Here we propose an extension to the ME architecture that is able to deal
with mappings overlapping in X space.
Another aspect of the results that is difficult to model using standard computational
architectures is memory consolidation. To account for this effect we introduce two
different learning rates (Alverez and Squire 1994). Some connections in the network
change faster, as a result of which they can serve as short-term memory. We also
introduce an off-line training phase (possibly corresponding to sleep) in which random inputs are generated, the part of the network containing the fast connections
is used to produce a target output, and the resulting input-output pair is used to
train the slowly changing connections. During the offline phase the faster changing
connections are fixed, after that they are randomized.
4.1
Modified Mixture of Experts
The ME model assumes that Y is generated from X by one of N processes
(Wl, ... , WN) and therefore the likelihood function is:
L(8IXt, Yt) = P(Ytlxt, 8) =
I: gi t P(YtIWi, Xt, 8)
i
g/ = P(WiIXt, 8),
where 8 represents the parameters of the model and gi is the prior probability. We
want to use the posterior probability P(Wi IXt, Yt) instead, because the processes
(different force fields) are separable in XY space, but not in X space. If we want to
implement an on-line controller such a term is not available, because at time t, Yt
is still unknown (the task of the controller is to produce it). We could approximate
P(WiIXt, Yt) with P(WdXt-l. Yt-!) , because dynamic conditions do not change very
often. Now the gating network (which computes P(Wdxt) in ME) is going to select
expert i based on the previous XY pair. This approach would obviously lead to a
single large error at the moment when the force field changes, but so will any model
using only Xt to compute Yt. In fact such an error seems to be consistent with our
psychophysics data. Thus the learning rule is:
~8i = lIih/(Yt - J.'i(Xt, 8i))
gI.t -_ h.t-1
I
h .t
I
h:
= P(W~I
I
~ ~ 8) =
XII
y"
gi t P(YtIWi , Xt, 8)
'"
L..tj g,.tp(Yt IW.
"Xt, 8)'
where
is the posterior probability and J.'i is the output of exper i. J.'i is a linear
function of the inputs. In our model we used 4 experts. In order to model the process
of consolidation, we gave one expert a learning rate that was 10 times higher than
the learning rate of the other 3 experts.
Catastrophic Interference in Human Motor Learning
25
4.2 The Model
We simulated the dynamics of a 2-joint planar arm similar to the one used in our
previous work (Shadmehr and Mussa-Ivaldi, 1994). The torque applied to the arm
at every point in time is the sum of the outputs of a fixed controller, a PO controller,
and an adaptive controller with the architecture described above.
The fixed controller maps (), iJ (current state), and jj (desired next state) into a
torque T. The mapping is exact when no external forces are applied. The desired trajectories are minimum-jerk trajectories (Flash and Hogan 1985) sampled
at 100Hz. The desired trajectories are 10 cm long and last 0.5 seconds. The PO
controller is used to compensate for the troques produced by the force field while
the adaptive controller has not yet learned to do that. The adaptive part of the
controller consists of a mixture of 4 linear experts (whose initial output is small)
and a modified gating network described above. The system operates as follows:
(), iJ, jj are sent to the fixed controller, which outputs a torque Tl; the PO controller
outputs a torque T2 based on the current deviation from the desired joint position
and velocity; 8 terms describing the current state of the arm (and chosen to linearize
the mapping to be learned) are sent to the mixture model, which outputs a torque
T3; Te = Tl + T2 - T3 is applied to the plant as a control signal; the actual torque
T = Te + TJ is computed. The mixture model is trained to produce the torque TJ
resulting from the force field. In other words, the adaptive part of the controller
learns to compensate for the force field exerted by the environment.
The parameters of the mixture model are updated after every movement, so a
training pair (Xt, Yt) is actually a batch of 50 points. The input, Xt, consists of
the 8 terms describing the current state and the desired next state; the output, Yt,
is the torque vector that the force field produces.The compensatory torques for a
complete movement are computed before the movement starts. The only processing
done during the movement is the computations necessary for the PO controller.
4.3 Results
4.3.1 Negative Transfer
When the network was given two successive, incompatable mappings to learn (this
corresponds to learning to move in two opposite force fields), the resulting performance very much resembled that of our human subjects. The performance in the
second mapping was much poorer than that in the first mapping. The fast-learning
expert changed its weights to learn both tasks. Since the two tasks involved anticorrelated maps, the fast expert's weights after learning the first mapping were very
inappropriate for the second task, leading to the observed negative transfer.
4.3.2 Catastrophic Interference
When the network was trained on two successive, opposite force fields, with no
consolidation occurring between the two training sessions, the learning in second
training session overwrote the learning that occurred during the first training session
(fig 3A). Since the expert with the fast-changing weights attempted to learn both
mappings, this catastrophic interference is not unexpected.
4.3.3 Consolidation
When the network was allowed to undergo consolidation between learning the first
and the second force field, the network no longer suffered from catastrophic inter-
26
Tom Braslzers-Krug, Reza Slzadmelzr, Emanuel Todorov
A
--:="
D.65~_ _ _ _~_ _ _
o
10
20
Figure 3: A: Learning curves for the ME architecture. Dark line: curve when first learning
Field A. light line: curve when given Field A a second time, after learning FIeld B (no
consolidation allowed between learning Field A and Field B). Note lack of retention of
Field A. B: Learning curves for the same architecture in Field A before and after learning
Field B. Consolidations was allowed between learning Field A and Field B.
ference (fig 3B). The learning that had initially resided in the fast-learning expert
was transfered to one of the slower-learning networks. Thus, when the expert with
the fast-changing connections learned the second mapping, the original learning was
no longer destroyed. In addition, when the network was allowed to consolidate the
second force field, a different slow-learning expert stored the second mapping. In
this way, the network stored multiple maps in long-term memory.
5
Conclusions
We have presented psychophysical evidence for catastrophic interference. We have
also shown results that suggest that motor learning can undergo a process of consolidation. By adding distinct fast- and slow-changing weights to a mixture of experts
architecture, we were able to account for these psychophysical findings. We plan to
investigate further the neural correlates of consolidation using both brain imaging
in humans and electrophysiological studies in primates.
References
P. Alverez and L. Squire. (1994) Memory consolidation and the medial temporal love: a
simple network model, PNAS91:15:7041-7045. T. Brashers-Krug, et al. (1994) Temporal
aspects of human motor learning, Soc. Neurosci. Abstract in press.
T. Flash and N. Hogan. (1985) The coordination of arm movements: an experimentally
confirmed mathematical model. J. Neurosci. 5:1688-1703.
R. Jacobs and M. Jordan. (1993) Learning piecewise control strategies in a modular neural
network architecture. IEEE Trans. on Systems, Man and Cyber. 23:2: 337-345.
I. Jenkins, et al. (1994) Motor sequence learning: a study with positron emission tomography, J. Neurosci.14:3775-3790.
M. Jordan and R. Jacobs. (1994) Hierarchical mixture of experts and the EM algorithm,
Neural Computation 6:2: 181-214.
M. Jordan and D. Rumelhart. (1992) Forward models: supervised learning with a distal
teacher Cognitive Sci. 16:307-354
A. Karni and D. Sagi. (1993) Nature 365:250.
R. Shadmehr and F . Mussa-Ivaldi. (1994) Adaptive representation of dynamics during
learning of a motor task, J. Neurosci.14:5:3208-3224.
R. Shadmehr, T. Brashers-Krug, F. Mussa-Ivaldi, (1995) Interference in learning internal
models of inverse dynamics in humans, Adv Neural Inform Proc Systvol 7 , in press
| 973 |@word seems:2 open:1 eng:1 jacob:4 fifteen:1 ld:1 moment:3 initial:4 ivaldi:5 series:2 practiced:3 current:6 yet:1 must:2 john:1 subsequent:1 plasticity:1 motor:16 displace:1 wanted:1 plot:1 medial:1 half:1 cue:1 manipulandum:4 nervous:1 positron:1 short:1 location:1 successive:2 five:1 mathematical:1 along:1 direct:1 become:2 acheived:1 consists:2 introduce:2 manner:1 inter:1 behavior:2 love:1 brain:2 torque:10 actual:1 inappropriate:1 circuit:2 what:2 viscous:3 consolidated:1 cm:1 developed:1 finding:2 temporal:2 every:2 control:16 partitioning:1 appear:1 before:8 retention:4 sagi:2 consequence:1 krug:7 path:1 approximately:2 tbk:1 suggests:2 practice:5 implement:1 jhu:1 significantly:3 pre:1 word:2 suggest:2 cannot:1 storage:3 context:2 map:6 imposed:1 yt:10 convex:2 immediately:4 rule:1 handle:2 notion:1 overwrite:1 coordinate:1 updated:1 target:4 engage:1 exact:1 substrate:1 velocity:6 element:1 rumelhart:2 observed:1 adv:1 movement:10 counter:2 environment:3 convexity:2 dynamic:20 hogan:2 trained:2 solving:1 predictive:1 deliver:1 serve:1 po:4 joint:6 train:2 univ:1 distinct:1 fast:8 tell:1 whose:1 modular:3 ability:3 gi:4 transform:1 itself:1 obviously:1 sequence:1 sen:1 propose:1 causing:1 rapidly:1 poorly:1 adapts:1 produce:5 linearize:1 bme:1 measured:1 ij:3 minor:1 solves:2 soc:1 predicted:1 resemble:1 come:1 indicate:1 met:1 resided:1 direction:1 correct:2 human:10 preliminary:1 biological:1 extension:1 mm:1 mapping:15 scope:1 proc:1 iw:1 coordination:1 wl:1 mit:1 modified:2 emission:1 improvement:3 indicates:1 likelihood:1 baseline:3 dependent:4 initially:2 going:1 selects:1 flexible:1 ill:1 ference:1 plan:1 psychophysics:1 field:90 exerted:2 having:1 ted:1 never:2 represents:2 alter:1 t2:2 piecewise:1 resulted:3 fileds:2 mussa:5 phase:2 investigate:1 mixture:9 light:5 tj:3 poorer:1 necessary:1 experience:2 xy:2 old:1 circle:1 plotted:2 desired:7 minimal:1 increased:2 modeling:2 soft:1 tp:1 organizational:1 deviation:1 ixt:2 stored:2 teacher:1 twelve:2 randomized:1 retain:4 off:1 transfered:2 hopkins:1 recorded:1 containing:1 possibly:1 slowly:1 worse:1 cognitive:2 external:1 expert:19 leading:2 return:1 account:3 sec:1 squire:2 later:1 portion:1 start:1 recover:1 who:1 t3:2 plastic:1 produced:1 trajectory:8 confirmed:1 straight:1 explain:1 inform:1 email:1 sensorimotor:2 involved:1 static:1 sampled:1 emanuel:5 electrophysiological:1 actually:1 back:1 appears:1 higher:2 day:25 supervised:3 tom:5 planar:3 done:1 biomedical:1 correlation:1 hand:2 overlapping:1 lack:1 interfere:1 indicated:1 effect:3 laboratory:1 deal:1 distal:3 during:5 whereby:1 presenting:1 complete:1 meaning:1 novel:2 unlearn:1 reza:6 volume:2 discussed:1 occurred:1 functionally:1 significant:1 cambridge:1 ai:3 curl:3 session:3 had:5 immune:2 robot:1 resistant:1 longer:4 similarity:1 operating:1 something:1 posterior:2 reverse:1 store:1 certain:1 minimum:1 employed:1 determine:1 paradigm:3 clockwise:4 signal:5 multiple:2 desirable:1 smooth:1 faster:2 adapt:1 compensate:2 long:3 retrieval:1 paired:2 prediction:1 controller:19 circumstance:2 physically:1 represent:1 addition:1 want:2 baltimore:1 interval:1 suffered:1 undo:1 subject:35 undergo:3 hz:1 sent:2 cyber:1 jordan:6 feedforward:2 wn:1 destroyed:1 todorov:5 affect:1 jerk:1 gave:1 architecture:9 opposite:2 shift:1 whether:5 returned:1 cause:1 jj:3 action:1 dark:5 ten:1 tomography:1 millisecond:1 xii:1 group:20 key:1 monitor:1 changing:6 imaging:1 sum:1 inverse:8 angle:1 consolidate:1 comparable:1 followed:1 sleep:1 activity:1 precisely:1 aspect:2 performing:1 separable:1 em:1 wi:1 modification:1 primate:1 ytlxt:1 interference:17 previously:1 emo:1 describing:2 needed:1 know:1 whichever:1 end:2 available:1 jenkins:2 brashers:6 apply:1 limb:3 observe:1 hierarchical:1 appropriate:2 batch:1 slower:1 original:1 assumes:2 perturb:1 psychophysical:6 move:9 question:3 already:1 depart:1 strategy:1 md:1 separate:4 link:1 simulated:1 entity:1 capacity:2 sci:1 me:5 extent:1 modeled:1 retained:1 difficult:2 setup:1 holding:2 negative:6 unknown:2 perform:2 anticorrelated:1 anti:1 displayed:1 situation:1 interacting:1 introduced:1 pair:4 required:1 connection:6 compensatory:1 elapsed:1 learned:19 hour:2 trans:1 able:5 suggested:1 beyond:1 pattern:1 challenge:1 memory:4 suitable:1 force:38 arm:9 prior:1 literature:1 fully:1 plant:4 degree:1 consistent:1 heavy:1 row:2 changed:1 consolidation:19 placed:1 last:1 free:1 offline:1 guide:1 explaining:1 benefit:1 karni:2 curve:23 world:1 computes:1 sensory:1 forward:3 adaptive:5 correlate:1 approximate:1 skill:1 observable:1 active:2 decides:1 assumed:1 scratch:1 learn:12 transfer:6 nature:1 exper:1 protocol:2 did:1 neurosci:4 arrow:1 decrement:1 allowed:8 fig:8 referred:1 tl:2 screen:1 slow:3 lc:1 position:2 explicit:1 ib:1 learns:2 minute:5 load:4 xt:7 resembled:1 gating:3 explored:1 evidence:8 essential:1 incorporating:1 adding:1 effectively:1 magnitude:1 te:2 occurring:1 cartesian:1 demand:2 cursor:1 suited:1 backpropagate:1 depicted:1 explore:3 likely:1 forming:2 unexpected:1 vulnerable:1 corresponds:1 ma:1 identity:2 acceleration:1 flash:2 erased:1 man:1 change:7 experimentally:1 muscular:1 except:1 operates:2 shadmehr:10 called:2 pas:1 catastrophic:16 experimental:13 la:1 attempted:1 select:2 internal:3 people:1 dept:2 tested:4 phenomenon:2 correlated:1 |
7,214 | 974 | Model of a Biological Neuron as a Temporal
Neural Network
Sean D. Murphy and Edward W. Kairiss
Interdepartmental Neuroscience Program, Department of Psychology,
and The Center for Theoretical and Applied Neuroscience,
Yale University,
Box 208205, New Haven, CT 06520
Abstract
A biological neuron can be viewed as a device that maps a multidimensional temporal event signal (dendritic postsynaptic activations) into a
unidimensional temporal event signal (action potentials). We have
designed a network, the Spatio-Temporal Event Mapping (STEM)
architecture, which can learn to perform this mapping for arbitrary biophysical models of neurons. Such a network appropriately trained,
called a STEM cell, can be used in place of a conventional compartmental model in simulations where only the transfer function is important,
such as network simulations. The STEM cell offers advantages over
compartmental models in terms of computational efficiency, analytical
tractabili1ty, and as a framework for VLSI implementations of biological neurons.
1 INTRODUCTION
Discovery of the mechanisms by which the mammalian cerebral cortex processes and
stores information is the greatest remaining challenge in the brain sciences. Numerous
modeling studies have attempted to describe cortical information processing in frameworks as varied as holography, statistical physics, mass action, and nonlinear dynamics.
Yet, despite these theoretical studies and extensive experimental efforts, the functional
architecture of the cortex and its implementation by cortical neurons are largely a mystery.
Our view is that the most promising approach involves the study of computational models
with the following key properties: (1) Networks consist of large (> 103 ) numbers of neurons; (2) neurons are connected by modifiable synapses; and (3) the neurons themselves
possess biologically-realistic dynamics.
Property (1) arises from extensive experimental observations that information processing
and storage is distributed over many neurons. Cortical networks are also characterized by
sparse connectivity: the probability that any two local cortical neurons are synaptically
connected is typically less than 0.1. These and other observations suggest that key features
of cortical dynamics may not be apparent unless large, sparsely-connected networks are
studied.
Property (2) is suggested by the accumulated evidence that (a) memory formation is subserved by use-dependent synaptic modification, and (b) Hebbian synaptic plasticity is
present in many areas of the brain thought to be important for memory. It is also well
known that artificial networks composed of elements that are connected by Hebb-like synapses have powerful computational properties.
86
Sean D. Murphy, Edward W. Kairiss
Property (3) is based on the assumption that biological neurons are computationally more
complex than. for example. the processing elements that compose artificial (connectionist)
neural networks. Although it has been difficult to infer the computational function of cortical neurons directly from experimental data, models of neurons that explicitly incorporate biophysical components (e.g. neuronal geometry, channel kinetics) suggest a
complex, highly non-linear dynamical transfer function. Since the "testability" of a model
depends on the ability to make predictions in terms of empirically measurable single-neuron firing behavior, a biologically-realistic nodal element is necessary in the network
model.
Biological network models with the above properties (e.g. Wilson & Bower, 1992; Traub
and Wong, 1992) have been handicapped by the computationally expensive single-neuron
representation. These "compartmental" models incorporate the neuron's morphology and
membrane biophysics as a large (102 _104) set of coupled, non-linear differential equations. The resulting system is often stiff and requires higher-order numerical methods and
small time-steps for accurate solution. Although the result is a realistic approximation of
neuronal dynamics, the computational burden precludes exhaustive study of large networks for functionality such as learning and memory.
The present study is an effort to develop a computationally efficient representation of a
single neuron that does not compromise the biological dynamical behavior. We take the
position that the "dynamical transfer function" of a neuron is essential to its computational
abstraction, but that the underlying molecular implementation need not be explicitly represented unless it is a target of analysis. We propose that a biological neuron can be viewed
as a device that performs a mapping from multidimensional spatio-temporal (synaptic)
events to unidimensional temporal events (action potentials). This computational abstraction will be called a Spatio-Temporal Event Mapping (STEM) cell. We describe the architecture of the neural net that implements the neural transfer function, and the training
procedure required to develop realistic dynamics. Finally, we discuss our preliminary
analyses of the performance of the model when compared with the full biophysical representation.
2 STEM ARCHITECTURE
The architecture of the STEM cell is similar to that found in neural nets for temporal
sequence processing (e.g. review by Mozer, in press). In general, these networks have 2
components: (1) a short-term memory mechanism that acts as a preprocessor for (2) a nonlinear feedforward network. For example, de Vries & Principe (1992) describe the utility
of the gamma net, a real-time neural net for temporal processing, in time series prediction.
The preprocessor in the gamma net is the gamma memory structure, implemented as a network of adaptive dispersive elements (de Vries & Principe, 1991). The preprocessor in our
model (the "tau layer", described below) is somewhat simpler, and is inspired by the temporal dynamics of membrane conductances found in biological neurons.
The STEM architecture (diagrammed in Figure 1) works by building up a vectorial representation of the state of the neuron as it continuously receives incoming synaptic activations, and then labeling that vector space as either "FIRE" or "DON'T FIRE". This is
accomplished with the use of four major components: (1) TAU LAYER: a layer of nodes
that continuously maps incoming synaptic activations into a finite-dimensional vector
space (2) FEEDBACK TAU NODE: a node that maintains a vectorial representation of
the past activity of the cell itself (3) MLP: a multilayer perceptron that functions as a nonlinear spatial mapping network that performs the "FIRE" / "NO-FIRE" labeling on the tau
layer output (4) OUTPUT FILTER: this adds a refractory period and threshold to the MLP
output that contrains the format of the output to be discrete-time events.
Model of Biological Neuron as a Temporal Neural Network
87
input (spike train at each synapse)
"
V
Ispatio-temporalto spatial projection layer !_
u
l
nonlinear spatial mapping network
Q
-
I
spike train output
I output processor :
Figure 1: Information Flow in The STEM Cell
The tau layer (Fig. 2) consists of N + 1 tau nodes, where N is the number of synapses on
the cell, and the extra node is used for feedback. Each tau node consists of M tau units.
Each tau unit within a single tau node receives an identical input signal. Each tau unit
within a tau node calculates a second-order rise-and-decay function with unique time constants. The tau units within a tau node translate arbitrary temporal events into a vector
form, with each tau-unit corresponding to a different vector component. Taken as a whole,
all of the tau unit outputs of the tau node layer comprise a high-dimensional vector that
represents the overall state of the neuron. Functionally, the tau layer approximates a oneto-one mapping between the spatio-temporal input and the tau-unit vector space.
The output of each tau unit in the tau layer is fed into the input layer of a multilayer perceptron (MLP) which, as will be explained in the next section, has been trained to label the
tau-layer vector as either FIRE or NO-FIRE. The output of the MLP is then fed into an
output filter with a refractory period and threshold. The STEM architecture is illustrated in
Fig. 3.
(A)
(afferents)
(feedback from action potentials)
synapse #
1,* ~ 3~?" N~ I ~+TAU_NODE,a~r
(all outputs go to MLP)
p~esynaptic
mput
(B)
TAU-UNIT DYNAMICS
(C)
d '
,
xiI
-xl = a)-dt I
'to
I
_d T,;
, TiI
.\ = xl-dt I
I
't.
I
ai(t)
I
TAU NODE
II
TAU UNITS
x~(~
tM....r"-T~(t~ .~~
Tlll_ _
Figure 2: Tau Layer Schematic. (A) the tau layer has an afferent section, with N tau
nodes, and a single-node feedback section. (B) Each tau node contains M tau units, and
therefore has 1 input and M outputs (C) Each of the M tau units in a tau node has a rise
and decay function with different constants. The equations are given for the ith tau unit of
the jth tau node. a is input activity, x an internal state variable, and T the output.
Sean D. Murphy, Edward W. Kairiss
88
single tau node from output with B
internal tau units
afferent inputs (axons)
&
/
IaU node layer
sN
B units
in 1st layer
from feedback
hidden layer of MLP
NxM units in axon subset of
1st layer of multilayer perceptron
single output unit of MLP
Schematic of output filter processin~
MLP
x
',))f\",AU M -n-;f-
output filter
I I I
= full forward connections
between layers
output axon
Figure 3: STEM Architecture. Afferent activity enters the tau layer, where it is converted into a vectorial representation of past spaiotemporal activity. The MLP maps this
vector into a FlRE/NO-FIRE output unit, the continuous value of which is converted to
a discrete event signal by the refractory period and threshold of the output filter.
3 STEM TRAINING
There are six phases to training the STEM cell:
(1) Biology: anatomical and physiological data are collected on the cell to be modeled.
(2) Compartmental Model: a compartmental model of the cell is designed, typically with a
simulation environment such as GENESIS. As much biological detail as possible is incorporated into the model.
(3) Transfer Function Trials: many random input sequences are generated for the compartmental model. The firing response of the model is recorded for each input sequence.
(4) Samplin~ assi~nments; In the next step, sampling will need to be done on the affect of
the input sequences on the STEM tay layer. The timing of the sampling is calculated by
separating the response of the compartmental model on each trial into regions where no
spikes occur, and regions surrounding spikes. High-rate sampling times are determined for
spike regions, and lower rate times are determined for quiet regions.
(5) Tau layer trials: the identical input sequences applied to the compartmental model in
step #3 are applied to an isolated tau layer of the STEM cell. The spike events from the
compartmental model are used as input for the feedback node. For each input sequence,
the tay layer is sampled at the times calculated in step #4, and the vector is labeled as
FIRE or NO-FIRE (0 or 1).
(6) MLP training: conjugate-gradient and line-search methods are used to train the multilayer perceptron using the results of step #5 as training vectors.
Model of Biological Neuron as a Temporal Neural Network
89
Training is continued until a minimal performance level is reached, as determined by comparing the response of the STEM cell to the original compartmental model on novel input
sequences.
4 RESULTS
The STEM cell has been initially evaluated using Roger Traub's (1991) compartmental
model of a hippocampal CAl cell, implemented in GENESIS by Dave Beeman. This is a
relatively simple model structurally, with 19 compartments connected in a linear segment,
with the soma in the middle. Dynamically, however, it is one of the most accurate and
sophisticated models published, with on the order of 100 voltage- and Ca++ sensitive
membrane channel mechanisms. 94 synapses were placed on the model. Each synapse
recevied a random spike train with average frequency 10Hz during training. A diagram of
the model and the locations of synaptic input is given in Fig. 4.
Inputs going to a single compartment were treated as members of a common synapse, so
there were a total of 13 tau nodes, with 5 tau units per node, for a total of 65 tau units, plus
5 additional units from the feedback tau node. These fed into 70 units in the input layer of
the MLP. Two STEM cells were trained, one on a passive shell of the CAl cell, and the
other with all of the membrane channels included. Both used 70 units in the hidden layer
1200m
880m
6mlfllll_m
14m
mwmiil
8m
received 8 synaptic inputs
Fig. 4 Structure of Traub's CAl cell
compartmental model
-42.
-+4.
-44'.,
-48.
-150.
-!5Z.
-64.
-116.
-ee.
0.5
1.0
1.5
2.0
2.5
0.5
1.0
1.5
20
2.5
STEM cell
10.0<H'"----t------+-----t----+-----+-------,.
B.
4.
2.
Figure 5: CAl Passive Cell Output. The somatic voltage for the passive compartmental
model and the corresponding output-filtered spike events are given in the upper graph.
The lower graph shows the repsonse of the STEM cell to the same input. The upper
trace of the lower graph is the output filter response, the lower trace is the raw output of
the MLP. Horizontal axes in seconds. Vertical axis: top, m V, and bottom, arbitrary units.
90
Sean D. Murphy, Edward W. Kairiss
compartmental model
10, 00-0, 00-
-10, 00-
-20,,00-30,.00-40,.00-
-eo,,00-
r---
-60,,00-
0.5
1.0
1.5
r--2.0
r--
I'::
2.5
STEM cell
10,i'lO"!'"----t-----+-----t-----t------t-------r
8,
0.5
1.0
1.5
2.0
2.5
Figure 6: Response of Active CAl Cell, The upper graph is the somatic voltage of the
CA 1 cell in response to a random input The lower graph is the response of the STEM
cell to the same input. The upper trace of the STEM graph is the output filter response,
and the lower trace is the raw MLP output.Horizontal axes in seconds, Vertical axis:
top, m V, and bottom, arbitrary units,
of the MLP, Comparisons of the compartmental model vs. STEM are shown in Fig. 5 and
, for the passive and active models, respectively, Fig. 6.
For the active cell model, the STEM cell was approximately 10 times faster, and used 10%
of the memory as the compartmental model.
5 DISCUSSION
The STEM cell is a general spatio-temporal node architecture and is similar to many context networks that have been previously developed. Its role as a single node in a large
meta-network is unexplored, albeit interesting because of its capacity to mimic the transfer
functions of biological neurons. Between the complexity range of connectionist networks
to biological networks, there may be a multitude of useful computational schemes for
solving different types of problems. The STEM cell is an attempt to efficiently capture
some elements of biological neural function in a form that can be scaled along this range.
Characterization of the computational function of the neuron is a topic of considerable
interest and debate. STEM cells may be useful as a rough measure of how complex the
transfer function of a given biophysical model is. For example, it might be able to answer
the question: Which instantiates a more sophisticated nonlinear spatio-temporal map, a
single-compartment cell with complex somatic Ca++ dynamics, or a cell with active Na+
channels in a complex dendritic tree?
STEM architecture may also be interesting for theoretical and applied ANN research as a
connectionist representation of a biological neuron. The expanding body of work on
focused network architectures (Mozer, 1989; Stometta, 1988) may be an avenue towards
the formalization of biological neural transfer functions. Because a VLSI implementation
Model of Biological Neuron as a Temporal Neural Network
91
of a STEM cell could be reprogrammed on the fly to assume the transfer function of any
pre-trained biophysically-modeled cell, a VLSI STEM network chip is a more versatile
approach to artifical implementations of biological neurons than reconstructing compartmental models in VLSI.
Our future plans include using networks with delay lines and hebbian learning rules, both
of which the STEM architecture is directly suited for, to investigate the capacity for
STEM networks to perform real-time dynamic pattern tracking. The present implementation of the STEM cell is by no means an optimal one. We are experimenting with alternative components for the MLP, such as modular recurrent networks.
Acknowledgements
This work was supported by the Yale Center for Theoretical and Applied Neuroscience.
References
de Vries, B. and Principe, J.C. (1991) A theory for neural networks with time delays. In
R.P. Lippmann, J. Moody, & D.S. Touretzlcy (Eds.), Advances in Neural Information Processing Systems 3 (pp. 162-168). San Mateo, CA: Morgan Kaufmann.
de Vries, B. and Principe, J.C. (1992) The gamma model - A new neural net model for
temporal processing. Neural Networks, 5, 565-576.
Mozer, M.C. (1989). A focused back-propagation algorithm for temporal pattern recognition, Complex Systems, 3, 349-381.
Mozer, M.C. (in press) Neural net architectures for temporal sequence processing. In A.
Weigend & N. Gershenfeld (Eds.), Predicting the Future and Understanding the Past.
Redwood City, CA: Addison-Wesley.
Stornetta, W.S., Hogg, T., & Huberman, B.A. (1988). A dynamical approach to temporal
pattern processing. In Anderson D.Z. (Ed.), Neural Information Processing Systems, 750759.
Traub, R.D. and Wong, R. K. S. (1991). Neuronal Networks of the Hippocampus. Cambridge: Cambridge University Press.
Wilson, M. and Bower, J.M. (1992) Cortical oscillations and temporal interactions in a
computer simulation of piriform cortex. J. Neurophysiol. 67:981-95.
| 974 |@word trial:3 middle:1 hippocampus:1 simulation:4 versatile:1 series:1 contains:1 past:3 comparing:1 activation:3 yet:1 numerical:1 realistic:4 plasticity:1 designed:2 v:1 device:2 ith:1 short:1 oneto:1 filtered:1 characterization:1 node:24 location:1 simpler:1 nodal:1 along:1 differential:1 consists:2 compose:1 behavior:2 themselves:1 morphology:1 brain:2 inspired:1 underlying:1 mass:1 developed:1 temporal:22 unexplored:1 multidimensional:2 act:1 scaled:1 unit:25 local:1 timing:1 despite:1 firing:2 approximately:1 might:1 plus:1 au:1 studied:1 mateo:1 dynamically:1 range:2 unique:1 implement:1 procedure:1 area:1 thought:1 projection:1 pre:1 suggest:2 kairiss:4 cal:5 storage:1 context:1 wong:2 conventional:1 map:4 measurable:1 center:2 go:1 focused:2 rule:1 continued:1 target:1 element:5 expensive:1 recognition:1 mammalian:1 sparsely:1 labeled:1 bottom:2 role:1 fly:1 enters:1 capture:1 region:4 connected:5 mozer:4 environment:1 complexity:1 touretzlcy:1 testability:1 dynamic:9 diagrammed:1 trained:4 solving:1 segment:1 compromise:1 efficiency:1 neurophysiol:1 chip:1 represented:1 surrounding:1 train:4 describe:3 artificial:2 labeling:2 formation:1 exhaustive:1 apparent:1 modular:1 compartmental:17 precludes:1 ability:1 itself:1 sequence:8 advantage:1 biophysical:4 analytical:1 net:7 propose:1 interaction:1 translate:1 develop:2 recurrent:1 received:1 edward:4 implemented:2 involves:1 functionality:1 filter:7 preliminary:1 biological:18 dendritic:2 kinetics:1 mapping:7 major:1 label:1 sensitive:1 city:1 rough:1 voltage:3 wilson:2 ax:2 experimenting:1 dependent:1 abstraction:2 accumulated:1 typically:2 initially:1 hidden:2 vlsi:4 going:1 overall:1 plan:1 spatial:3 comprise:1 sampling:3 identical:2 represents:1 biology:1 mimic:1 future:2 connectionist:3 haven:1 composed:1 gamma:4 murphy:4 geometry:1 phase:1 fire:9 attempt:1 conductance:1 mlp:15 interest:1 highly:1 investigate:1 traub:4 accurate:2 necessary:1 unless:2 tree:1 isolated:1 theoretical:4 minimal:1 modeling:1 subset:1 delay:2 answer:1 st:2 physic:1 continuously:2 moody:1 na:1 connectivity:1 recorded:1 potential:3 tii:1 de:4 converted:2 explicitly:2 afferent:4 depends:1 view:1 reached:1 maintains:1 compartment:3 kaufmann:1 largely:1 efficiently:1 raw:2 biophysically:1 dave:1 processor:1 published:1 synapsis:4 synaptic:7 ed:3 frequency:1 pp:1 sampled:1 dispersive:1 sean:4 sophisticated:2 back:1 wesley:1 higher:1 dt:2 response:8 synapse:4 done:1 box:1 evaluated:1 anderson:1 roger:1 until:1 receives:2 horizontal:2 nonlinear:5 propagation:1 holography:1 building:1 illustrated:1 during:1 hippocampal:1 performs:2 passive:4 novel:1 common:1 functional:1 empirically:1 refractory:3 cerebral:1 approximates:1 functionally:1 cambridge:2 ai:1 hogg:1 cortex:3 add:1 stiff:1 store:1 meta:1 accomplished:1 morgan:1 additional:1 somewhat:1 eo:1 period:3 signal:4 ii:1 full:2 infer:1 stem:33 hebbian:2 faster:1 characterized:1 offer:1 molecular:1 biophysics:1 schematic:2 prediction:2 calculates:1 multilayer:4 synaptically:1 cell:34 diagram:1 appropriately:1 extra:1 posse:1 hz:1 member:1 flow:1 ee:1 feedforward:1 affect:1 psychology:1 architecture:13 tm:1 unidimensional:2 avenue:1 six:1 utility:1 effort:2 action:4 useful:2 neuroscience:3 per:1 modifiable:1 anatomical:1 xii:1 discrete:2 key:2 four:1 soma:1 threshold:3 gershenfeld:1 graph:6 weigend:1 mystery:1 powerful:1 place:1 oscillation:1 layer:25 ct:1 yale:2 activity:4 occur:1 vectorial:3 reprogrammed:1 format:1 relatively:1 department:1 instantiates:1 conjugate:1 membrane:4 reconstructing:1 postsynaptic:1 biologically:2 modification:1 explained:1 taken:1 computationally:3 equation:2 nxm:1 previously:1 discus:1 mechanism:3 addison:1 fed:3 alternative:1 original:1 top:2 remaining:1 include:1 question:1 spike:8 gradient:1 quiet:1 assi:1 separating:1 capacity:2 topic:1 collected:1 modeled:2 piriform:1 difficult:1 debate:1 trace:4 processin:1 rise:2 implementation:6 perform:2 upper:4 vertical:2 neuron:29 observation:2 finite:1 incorporated:1 incorporate:2 genesis:2 varied:1 somatic:3 redwood:1 arbitrary:4 required:1 extensive:2 connection:1 able:1 suggested:1 dynamical:4 below:1 pattern:3 handicapped:1 challenge:1 program:1 memory:6 tau:43 greatest:1 event:11 treated:1 predicting:1 scheme:1 numerous:1 axis:2 coupled:1 sn:1 review:1 understanding:1 discovery:1 acknowledgement:1 interesting:2 lo:1 placed:1 supported:1 jth:1 perceptron:4 sparse:1 distributed:1 feedback:7 calculated:2 cortical:7 forward:1 adaptive:1 san:1 lippmann:1 active:4 incoming:2 spatio:6 don:1 continuous:1 search:1 subserved:1 promising:1 learn:1 transfer:9 channel:4 ca:5 expanding:1 complex:6 whole:1 stornetta:1 body:1 neuronal:3 fig:6 hebb:1 axon:3 formalization:1 structurally:1 position:1 xl:2 bower:2 preprocessor:3 decay:2 physiological:1 multitude:1 evidence:1 consist:1 burden:1 essential:1 albeit:1 vries:4 suited:1 tracking:1 shell:1 viewed:2 ann:1 towards:1 considerable:1 included:1 determined:3 tay:2 huberman:1 called:2 total:2 experimental:3 attempted:1 principe:4 internal:2 mput:1 arises:1 artifical:1 |
7,215 | 975 | Dynamic Cell Structures
Jorg Bruske and Gerald Sommer
Department of Cognitive Systems
Christian Albrechts University at Kiel
24105 Kiel- Germany
Abstract
Dynamic Cell Structures (DCS) represent a family of artificial neural
architectures suited both for unsupervised and supervised learning.
They belong to the recently [Martinetz94] introduced class of Topology
Representing Networks (TRN) which build perlectly topology preserving feature maps. DCS empI'oy a modified Kohonen learning rule
in conjunction with competitive Hebbian learning. The Kohonen type
learning rule serves to adjust the synaptic weight vectors while Hebbian
learning establishes a dynamic lateral connection structure between
the units reflecting the topology of the feature manifold. In case of supervised learning, i.e. function approximation, each neural unit implements
a Radial Basis Function, and an additional layer of linear output units
adjusts according to a delta-rule. DCS is the first RBF-based approximation scheme attempting to concurrently learn and utilize a perfectly topology preserving map for improved performance.
Simulations on a selection of CMU-Benchmarks indicate that the DCS
idea applied to the Growing Cell Structure algorithm [Fritzke93] leads
to an efficient and elegant algorithm that can beat conventional models
on similar tasks.
1
Introduction
The quest for smallest topology preserving maps motivated the introduction of growing
feature maps like Fritzke's Growing Cell Structures (GCS). In GCS, see [Fritzke93] for details, one starts with a k-dimensional simplex of N = k+ 1 neural units and (k + 1) . kl2
lateral connections (edges). Growing of the network is performed such that after insertion
498
Jorg Bruske, Gerald Sommer
of a new unit the network consists solely of k dimensional simplices again. Thus, like Kohonen's SOM, GCS can only learn a perfectly topology preserving feature mapl if k
meets the actual dimension of the feature manifold. Assuming that the lateral connections
do reflect the actual topology the connections serve to define a neighborhood for a Kohonen
like adaptation of the synaptic vectors Wj and guide the insertion of new units. Insertion
happens incrementally and does not necessitate a retraining of the network. The principle
is to insert new neurons in such a way that the expected value of a certain local error measure, which Fritzke ca11s the resource, becomes equal for all neurons. For instance, the
number of times a neuron wins the competition, the sum of distances to stimuli for which
the neuron wins or the sum of errors in the neuron's output can all serve as a resource and
dramatically change the behavior of GCS. Using different error measures and guiding insertion by the lateral connections contributes much to the success of GCS.
The principle of DCS is to avoid any restriction of the topology of the network (lateral connection scheme between the neural units) but to concurrently learn and utilize a perfectly
topology preserving map. This is achieved by adapting the lateral connection structure according to a competitive Hebbian learning rule2 :
max{Yj'YpCij(t)} :
CIj(t+ 1) = {
0
aCjj (t)
Yj'Yj~Yk'YI V'(1S,k,IS,N)
: Cjj(t) <9
fIfI
otherwise,
till
:
(1)
where a, 0 < a < 1 is a forgettin?. constant, 9, 0 < 9 < 1 serves as a threshold for deleting
lateral connections, and Yj = R (j/v - Wj//) is the activation of the i-th unit with Wj as the
centre of its receptive field on presentatIOn of stimulus v. R(.) can be any positive continuously monotonically decreasing function. For batch learning with a training set T of fixed
size 111 , a = 17JJ9 is a good choice.
Since the isomorphic representation of the topology of the feature manifold M in the lateral
connection structure is central to performance, in many situations a DCS algorithm may be
the right choice. These situations are characterized by missing a priori knowledge of the
topology of the feature manifold M or a topology of M which cannot be readily mapped to
the existing models. Of course, if such a priori knowledge is available then models like
GCS or Kohonen's SOM allowing to incorporate such knowledge have an advantage, especially if training data are sparse.
Note that DCS algorithms can also aid in cluster analysis: In a perfectly topology preserving map clusters which are bounded by regions of P(v) =0 can be identified simply by a
connected component analysis. However, without prior knowledge about the feature manifold M it is in principal impossible to check for perfect topology preservation of S. Noise
in the input data may render perfect topology learning even more difficult. So what can perfect topology learning be used for? The answer is simply that for every s~t S of reference
vectors perfect topology learning yields maximum topology preservation with respect to
this set. And connected components with respect to the lateral connection structure C may
well serve as an initialization for postprocessing by hierarchical cluster algorithms.
1. We use the term "perfectly topology preserving feature map" in accordance with its rigorous
definition in [Martinetz93].
2. In his very recent and recommendable article [Martinetz94] the term Topology Representing
Network (TRN) is coined for any network employing competitive Hebbian learning for topology learning.
3. if topology preservation is measured by the topographic function as defined in [Villmann94].
Dynamic Cell Structures
499
The first neural algorithm attempting to learn perfectly topology preserving feature maps
is the Neural Gas algorithm ofT. Martinetz [Martinetz92]. However, unlike DCS the Neu?
ral Gas does not further exploit this information: In every step the Neural Gas computes
the k nearest neighbors to a given stimulus and, in the supervised case, employs a11 of them
for function approximation. DCS avoids this computational burden by utilizing the lateral
connection structure (topology) learned so far, and it restricts interpolation between activated units to the submanifold of the current stimulus.
Applying the principle ofDCS to Fritzke's GCS yields our DCS?GCS algorithm. This algorithm sticks very closely to the basic structure of its ancestor GCS except the predefined
k-dimensional simplex connection structure being replaced by perfect topology learning.
Besides the conceptual advantage of perfect topology learning, DCS?GCS does decrease
overhead (Fritzke has to handle quite sophisticated data structures in order to maintain the
k-dimensional simplex structure after insertion! deletion of units) and can be readily implemented on any serial computer.
2
Unsupervised DCS-GCS
The unsupervised DCS-GCS algorithm starts with initializing the network (graph) to two
neural units (vertices) n, and n2. Their weight vectors wI' WI (centres of receptive fields)
are set to points v I' v2 E M which are drawn from M according to P(v). They are connected
by a lateral connection of weight C l2 = C21 = I. Note that lateral connections in DCS
are always bidirectional and have symmetric weights.
Now the algorithm enters its outer loop which is repeated until some stopping criterium is
fulfi11ed. This stopping criterium could for instance be a test whether the quantization error
has already dropped below a predefined accuracy.
The inner loop is repeated for A. times. In off-line learning A. can be set to the number examples in the training set T. In this case, the inner loop just represents an epoch of training.
Within the inner loop, the algorithm first draws an input stimulus v E M from M according
to P(v) and then proceeds to calculate the two neural units which weight vectors are first
and second closest to v.
In the next step, the lateral connections between the neural units are modified according to
eq. (1), the competitive Hebbian learning rule. As has already been mentioned, in off-line
learning it is a good idea to set a =
va .
Now the weight vectors
Kohonen like fashion:
Wj
of the best matching unit and its neighbors are adjusted in a
.1.wbmu =
where
the
?8
neighborhood
(v - wbmu> and .1.Wj
N h U)
of
a
= ?Nh (v unit
wj ) ,
j
is
(2)
defined
by
NhU) = {il(Cjj*O,l~i~N)} .
The inner loop ends with updating the resource value of the best matching unit. The resource of a neuron is a local error measure attached to each neural unit. As has been pointed
out, one can choose alternative update functions corresponding to different error measures.
For our ~xperim~nts (section 2.1 and sectpn 3.1) we used the accumulated squared distance
to the stImulus, I.e . .1.'tbmu =
wbmull .
The outer loop now proceeds by adding a new neural unit r to the network. This unit is located in-between the unit I with largest resource value and its neighbor n with second largest resource value: 4
Ilv -
500
Jorg Bruske, Gerald Sommer
The exact location of its centre of receptive field wr is calculated according to the ratio of
the resource values 'tl , 'tn' and the resource values of units n and I are redistributed among
r, n and I:
Wr
= wI + yew n - wI) , 't r = ~'tn + ~'tr' 't l = 't[- ~'tr and't n = 't n - ~'tn?
(4)
This gives an estimate of the resource values if the new unit had been in the network right
from the start. Finally the lateral connections are changed,
Cr=C[
rn =0,
r
r =l,C rn =Crn =IandCr=C
n
(5)
connecting unit r to unit I and disconnecting n and I.
This heuristic guided by the lateral connection structure and the resource values promises
insertion of new units at good initial positions. It is responsible for the better performance
of DCS-GCS and GCS compared to algorithms which do not exploit the neighborhood relation between existing units.
?
The outer loop closes by decrementing the resource values of all units,
't i (t + 1) = ~'ti (t) , 1:::; i:::; N, where < ~ < 1 is a constant. This last step just avoids
is the natural choice.
overflow of the resource variables. For off-line learning, ~ =
2.1
?
Unsupervised DCS simulation results
Let us first turn to our simulation on artificial data. The training set T contains 2000 examples randomly drawn from a feature manifold M consisting of three squares, two of them
connected by a line. The development of our unsupervised DCS-GCS network is depicted
in Figure 1, with the initial situation of only two units shown in the upper left. Examples
are represented by small dots, the centres of receptive fields by small circles and the lateral
connections by lines connecting the circles. From left to right the network is examined after
0, 9 and 31 epochs of training (i.e. after insertion of 2, 11 and 33 neural units).
After 31 epochs the network has built a perfectly topology preserving map of M, the lateral
connection structure nicely reflecting the shape of M: Where Mis 2-dimensional the lateral
connection structure is 2-dimensional, and it is I-dimensional where M is I-dimensional.
Note, that a connected component analysis could recognize that the upper right square is
separated from the rest of M. The accumulated squared distance to stimuli served as the resource.
Th.e quantization error Eq = ! L Ilv - wbmu (v) 112 dropped from 100% (3 units) to 3% (33
umts).
n VE T
The second simulation deals with the two-spirals benchmark. Data were obtained by running the program "two-spirals" (provided by eMU) with parameters 5 (density) and 6.5
(spiral radius) resulting in a training set T of 962 examples. The data represent two distinct
spirals in the x-y-plane. Unsupervised DCS-GCS at work is shown in Figure 2, after insertion of 80, 154 and, finally, 196 units. With 196 units a perfectly topology preserving map
of M has emerged, and the two spirals are clearly separated. Note that the algorithm has
learned the separation in a totally unsupervised manner, i.e. not using the labels of the data
4. Fritzke inserts new units at a slightly different location, using not the neighbor with second
largest resource but the most distant neighbor.
Dynamic Cell Structures
501
Figure 1: Unsupervised DCS-GCS on artificial data
points (which are provided by CMU for supervised learning). Again, the accumulated
squared distance to stimuli served as the resource.
-;..---
------,--- -- - .
,
'-::---.:-:r~~
~.:~:.~.';{';';'; ~.:.;.! ~~~! ~I
Figure 2: Unsupervised learning of two spirals
3
Supervised DCS-GCS
In supervised DCS-GCS examples consist not only of an input vector v but also include an
additional teaching output vector u.
The supervised algorithm actually does work very similar to its unsupervised version except
? when a neural unit nj is inserted an output vector OJ will be attached to it with
OJ = u.
? the output y of the network is calculated as a weighted sum of the best matching unit's
output vector 0hmu and the output vectors of its neighbors OJ' i E Nh (bmu) ,
Y
= (~
~jE
{bmuuNh(hmu)}
a.o.)
,
I I
(6)
502
Jorg Brnske, Gerald Sommer
= I/(crllv-wiI12+ 1) is the activation of neuron i on stimulus v,
cr, cr> (1, representing the size of the receptive fields. In our simulations, the size of
where a?
receptive fields have been equal for all units.
? adaption of output vectors by the delta-rule: A simple delta-rule is employed to adjust
the output vectors of the best matching unit and its neighbors.
Most important, the approximation (classification) error can be used for resource updating.
This leads to insertion of new units in regions where the approximation error is worst, thus
promising to outperform dynamic algorithms which do not employ such a criterion for insertion. In our simulations we used the accumulated squared distance of calculated and
teaching output, ~tbmu = Ily - u11 2 .
3.1 Supervised DCS-GCS simulation results
We applied our supervised DCS-GCS algorithm to three CMU benchmarks, the supervised
two-spiral problem, the speaker independent vowel recognition problem and the sonar
mine! rock separation problem. 5
The
t~o
vE
~
spirals benchmark contains 194 examples, each consisting of an input vector
and a binary label indicating to which spiral the point belongs. The spirals can not
be linearly separated. The task is to train the examples until the learning system can produce the correct output for all of them and to record the time.
The decision regions learned by supervised DCS-GCS are depicted in Figure 3 after 110
and 135 epochs of training, where the classification error on the training set has dropped to
0%. Black indicates assignment to the fist, white assignment to the second spiral. The network and the examples are overlaid.
Figure 3: Supervised learning of two spirals
Results reported by others are 20000 epochs of Backprop for a MLP by Lang and Witbrok
[Lang89], 10000 epochs of Cross Entropy Backprop and 1700 epochs of Cascade-Correlation by Fahlman and Lebiere [Fahlman90] and 180 epochs of GCS training by Fritzke
[Fritzke93].
5. For details of simulation, parameters and additional statistics for all of the reported experiments the reader is refered to [Bruske94] which is also available viaftp.infomuztik.uni-kiel.de in
directory publkiellpublicationsffechnicalReportslPs.ZI as 1994tr03.ps.Z
Dynamic Cell Structures
503
The data for the speaker independent recognition of 11 vowels comprises a training set of
582 examples and a test set of 462 examples, see [Robinson89].
We obtained 65% correctly classified test samples with only 108 neural units in the DCS?
GCS network. This is superior to conventional models (including single and multi layer
perceptron, Kanerva Model, Radial Basis Functions, Gaussian Node Network, Square
Node Network and Nearest Neighbor) for which figures well below 57% have been reported by Robinson. It also qualitatively compares to GCS Gumps above the 60% margin), for
which Fritzke reports best classification results of 61 %(158 units) up to 67% (154 units) for
a 3-dim GCS. On the other hand, our best DCS?GCS used much fewer units. Note that
DCS?GCS did not rely on a pre-specified connection structure (but learned it!).
Our last simulation concerns a data set used by Gorman and Sejnowski in their study of
classification of sonar data, [Gorman88]. The training and the test set contain 104 examples
each.
Gorman and Sejnowski report their best results of 90.4% correctly classified test examples
for a standard BP network with 12 hidden units and 82.7% for a nearest neighbor classifier.
Supervised DCS?GCS reached a peak classification rate of 95% after only 88 epochs of
training.
4
Conclusion
We have introduced the idea ofRBF networks which concurrently learn and utilize perfectly topology preserving feature maps for adaptation and interpolation. This family of ANNs,
which we termed Dynamic Cell Structures, offers conceptual advantage compared to classical Kohonen type SOMs since the emerging lateral connection structure maximally preserves topology. We have discussed the DCS-GCS algorithm as an instance of DCS.
Compared to its ancestor GCS of Fritzke, this algorithm elegantly avoids computational
overhead for handling sophisticated data structures. If connection updates (eq.(I)) are restrJcted to the best matching unit and its neighbors, DCS has linear (serial) time complexity and thus may also be considersd as an improvement of Martinetz's Neural Gas idea7 .
Space complexity of DCS is 0 (N) in general and can be shown to become linear if the
feature manifold M is two dimensional. The simulations on CMU-Benchmarks indicate
that DCS indeed has practical relevance for classification and approximation.
Thus encouraged, we look forward to apply DCS at various sites in our active computer
vision project, including image compression by dynamic vector quantization, sensorimotor
maps for the oculomotor system and hand-eye coordination, cartography and associative
memories. A recent application can be found in [Bruske95] where a DCS network attempts
to learn a continous approximation of the Q-function in a reinforcement learning problem.
6. Here we refer to the serial time a DeS algorithm needs to process a single stimulus (including
response calculation and adaptation).
7. The serial time complexity of the Neural Gas is Q (N) ,approaching 0 (NlogN) for
k ~ N , k the number of nearest neighbors.
504
larg Bruske, Gerald Sommer
References
[Bruske94] J. Bruske and G. Sommer. Dynamic Cell Structures: Radial Basis Function Networks
with Perfect Topology Preservation. Inst. f. Inf. u. Prakt. Math. CAU zu Kiel. Technical Report 9403.
[Bruske9S] J. Bruske. I. Ahms and G. Sommer. Heuristic Q-Learning. submitted to ECML 95.
[Fahlman90] S.E. Fahlman. C.Lebiere. The Cascade-Correlation Learning Architecture. Advances
in Neural Information processing systems 2, Morgan Kaufman, San Mateo, pp.524-534.
[Fahlman93] S.E. Fahlman, CMU Benchmark Collection for Neural Net Learning Algorithms,
Carnegie Mellon Univ., School of Computer Science, machine-readable data repository. Pittsburgh.
[Fritzke92] B. Fritzke. Growing Cell Structures - a self organizing network in k dimensions, Artificial Neural Networks 2. I.Aleksander & J .Taylor eds .? North-Holland, Amsterdam, 1992.
[Fritzke93] B. Fritzke, Growing Cell Structures - a self organizing network for unsupervised and
supervised training, ICSI Berkeley, Technical Report, tr-93-026.
[Gorman88] R.B. Gorman and TJ. Sejnowski, Analysis of Hidden Units in a Layered Network
Trained to Classify Sonar Targets, Neural Networks, VoU. pp. 75-89
[Lang89] K.J. Lang & M.J. Witbrock, Learning to tell two spirals apart, Proc. of the 1988 Connectionist Models Summer School. Morgan Kaufmann, pp.52-59.
[Martinetz92] Thomas Martinetz, Selbstorganisierende neuronale Netzwerke zur Bewegungssteuerung, Dissertation. DIFKI-Verlag, 1992.
[Martinetz93] Thomas Martinetz, Competitive Hebbian Learning Rule Forms Perfectly Topology
Preserving Maps, Proc. of the ICANN 93. p.426-438, 1993.
[Martinetz94] Thomas Martinetz and Klaus Schulten, Topology Representing Networks, Neural
Networks. No.7, Vol. 3. pp. 505-522, 1994.
[Moody89] J.Moody, c.J. Darken. Fast Learning in Networks of Locally-Tuned Processing Units,
Neural Computation VoLl Num.2, Summer 1989.
[Robinson89] AJ. Robinson, Dynamic Error Propagation Networks, Cambridge Univ., Ph.D. thesis, Cambridge.
[Villmann94] T. Villmann and R. Der and T. Martinetz, A Novel Approach to Measure the Topology
Preservation of Feature Maps, Proc. of the ICANN 94, 1994.
| 975 |@word repository:1 version:1 compression:1 retraining:1 simulation:10 u11:1 tr:3 initial:2 contains:2 tuned:1 existing:2 current:1 nt:1 activation:2 lang:2 readily:2 distant:1 shape:1 christian:1 update:2 fewer:1 plane:1 directory:1 dissertation:1 record:1 num:1 math:1 node:2 location:2 kiel:4 albrechts:1 become:1 consists:1 overhead:2 manner:1 indeed:1 expected:1 behavior:1 growing:6 multi:1 decreasing:1 actual:2 totally:1 becomes:1 provided:2 project:1 bounded:1 what:1 kaufman:1 emerging:1 nj:1 berkeley:1 every:2 ti:1 classifier:1 stick:1 unit:46 positive:1 dropped:3 local:2 accordance:1 meet:1 solely:1 rule2:1 interpolation:2 black:1 initialization:1 mateo:1 examined:1 c21:1 practical:1 responsible:1 yj:4 cau:1 implement:1 nlogn:1 adapting:1 cascade:2 matching:5 pre:1 radial:3 cannot:1 close:1 selection:1 layered:1 impossible:1 applying:1 disconnecting:1 restriction:1 conventional:2 map:14 missing:1 rule:7 adjusts:1 utilizing:1 his:1 handle:1 target:1 exact:1 recognition:2 updating:2 located:1 inserted:1 initializing:1 enters:1 worst:1 calculate:1 wj:6 region:3 connected:5 decrease:1 icsi:1 yk:1 mentioned:1 insertion:10 complexity:3 mine:1 dynamic:11 gerald:5 trained:1 serve:3 basis:3 represented:1 various:1 train:1 separated:3 distinct:1 univ:2 fast:1 sejnowski:3 artificial:4 tell:1 klaus:1 neighborhood:3 quite:1 heuristic:2 emerged:1 otherwise:1 statistic:1 topographic:1 associative:1 advantage:3 net:1 rock:1 redistributed:1 adaptation:3 kohonen:7 loop:7 organizing:2 till:1 competition:1 cluster:3 p:1 produce:1 a11:1 perfect:7 measured:1 nearest:4 school:2 eq:3 implemented:1 indicate:2 guided:1 radius:1 closely:1 correct:1 backprop:2 adjusted:1 insert:2 overlaid:1 smallest:1 ilv:2 proc:3 label:2 coordination:1 largest:3 establishes:1 weighted:1 concurrently:3 clearly:1 always:1 gaussian:1 modified:2 avoid:1 cr:3 aleksander:1 conjunction:1 improvement:1 ily:1 ral:1 check:1 indicates:1 cartography:1 rigorous:1 dim:1 inst:1 stopping:2 accumulated:4 hidden:2 relation:1 ancestor:2 germany:1 among:1 classification:6 priori:2 development:1 equal:2 field:6 nicely:1 encouraged:1 represents:1 look:1 unsupervised:11 simplex:3 others:1 stimulus:10 report:4 connectionist:1 employ:2 randomly:1 preserve:1 recognize:1 ve:2 replaced:1 consisting:2 maintain:1 vowel:2 attempt:1 mlp:1 adjust:2 activated:1 tj:1 predefined:2 edge:1 taylor:1 circle:2 instance:3 classify:1 assignment:2 witbrock:1 vertex:1 submanifold:1 reported:3 answer:1 density:1 peak:1 off:3 connecting:2 continuously:1 moody:1 again:2 reflect:1 central:1 squared:4 thesis:1 choose:1 necessitate:1 cognitive:1 de:2 north:1 performed:1 reached:1 competitive:5 start:3 recommendable:1 il:1 square:3 accuracy:1 kaufmann:1 yield:2 yew:1 served:2 classified:2 submitted:1 anns:1 synaptic:2 neu:1 definition:1 ed:1 kl2:1 sensorimotor:1 pp:4 lebiere:2 mi:1 knowledge:4 sophisticated:2 actually:1 reflecting:2 bidirectional:1 supervised:14 response:1 improved:1 maximally:1 just:2 until:2 correlation:2 hand:2 propagation:1 incrementally:1 aj:1 acjj:1 contain:1 symmetric:1 deal:1 white:1 self:2 speaker:2 criterion:1 tn:3 postprocessing:1 image:1 novel:1 recently:1 superior:1 attached:2 nh:2 belong:1 discussed:1 refer:1 mellon:1 cambridge:2 pointed:1 teaching:2 centre:4 had:1 dot:1 vou:1 closest:1 recent:2 inf:1 belongs:1 apart:1 termed:1 verlag:1 certain:1 binary:1 success:1 yi:1 der:1 criterium:2 preserving:12 morgan:2 additional:3 employed:1 monotonically:1 preservation:5 fist:1 hebbian:6 technical:2 characterized:1 calculation:1 cjj:2 cross:1 offer:1 serial:4 va:1 basic:1 vision:1 cmu:5 represent:2 achieved:1 cell:11 zur:1 rest:1 unlike:1 martinetz:6 elegant:1 fritzke:10 spiral:13 zi:1 architecture:2 topology:34 perfectly:10 identified:1 inner:4 idea:3 approaching:1 whether:1 motivated:1 render:1 dramatically:1 locally:1 ph:1 outperform:1 restricts:1 delta:3 wr:2 correctly:2 carnegie:1 promise:1 vol:1 voll:1 threshold:1 drawn:2 utilize:3 graph:1 sum:3 family:2 reader:1 separation:2 draw:1 decision:1 layer:2 summer:2 larg:1 bp:1 attempting:2 department:1 trn:2 according:6 slightly:1 wi:4 happens:1 refered:1 kanerva:1 resource:16 turn:1 serf:2 end:1 available:2 apply:1 hierarchical:1 v2:1 batch:1 alternative:1 thomas:3 running:1 sommer:7 include:1 readable:1 exploit:2 coined:1 build:1 especially:1 overflow:1 classical:1 already:2 receptive:6 win:2 distance:5 mapped:1 lateral:19 outer:3 manifold:7 assuming:1 besides:1 ratio:1 difficult:1 cij:1 jorg:4 allowing:1 upper:2 neuron:7 darken:1 benchmark:6 gas:5 beat:1 ecml:1 situation:3 dc:35 gc:31 rn:2 introduced:2 specified:1 connection:23 continous:1 learned:4 deletion:1 emu:1 robinson:2 proceeds:2 below:2 oft:1 oculomotor:1 program:1 built:1 max:1 oj:3 including:3 deleting:1 memory:1 natural:1 rely:1 representing:4 scheme:2 eye:1 prior:1 epoch:9 l2:1 oy:1 article:1 principle:3 course:1 changed:1 last:2 fahlman:3 neuronale:1 guide:1 perceptron:1 neighbor:11 sparse:1 dimension:2 calculated:3 avoids:3 computes:1 forward:1 qualitatively:1 reinforcement:1 san:1 bmu:1 collection:1 employing:1 far:1 uni:1 active:1 conceptual:2 pittsburgh:1 decrementing:1 crn:1 sonar:3 promising:1 learn:6 contributes:1 som:3 elegantly:1 did:1 icann:2 linearly:1 noise:1 n2:1 bruske:6 repeated:2 site:1 je:1 tl:1 fashion:1 simplices:1 aid:1 position:1 guiding:1 comprises:1 schulten:1 zu:1 concern:1 burden:1 consist:1 quantization:3 adding:1 umts:1 margin:1 gorman:3 suited:1 entropy:1 depicted:2 simply:2 amsterdam:1 holland:1 adaption:1 presentation:1 rbf:1 change:1 except:2 principal:1 isomorphic:1 indicating:1 quest:1 relevance:1 incorporate:1 handling:1 |
7,216 | 976 | On-line Learning of Dichotomies
N. Barkai
H. S. Seung
AT&T Bell Laboratories
Murray Hill, NJ 07974
seungCphysics.att.com
Racah Institute of Physics
The Hebrew University
Jerusalem, Israel 91904
naamaCfiz.huji.ac.il
H. Sompolinsky
Racah Institute of Physics
The Hebrew University
Jerusalem, Israel 91904
and AT&T Bell Laboratories
haimCfiz.huji.ac.il
Abstract
The performance of on-line algorithms for learning dichotomies is studied. In on-line learning, the number of examples P is equivalent to the learning time, since each example is
presented only once. The learning curve, or generalization error as a function of P, depends
on the schedule at which the learning rate is lowered. For a target that is a perceptron rule,
the learning curve of the perceptron algorithm can decrease as fast as p- 1 , if the schedule is optimized. If the target is not realizable by a perceptron, the perceptron algorithm
does not generally converge to the solution with lowest generalization error. For the case
of unrealizability due to a simple output noise, we propose a new on-line algorithm for a
perceptron yielding a learning curve that can approach the optimal generalization error as
fast as p-l/2. We then generalize the perceptron algorithm to any class of thresholded
smooth functions learning a target from that class. For "well-behaved" input distributions,
if this algorithm converges to the optimal solution, its learning curve can decrease as fast
as p-l.
1
Introduction
Much work on the theory of learning from examples has focused on batch learning, in which the learner is
given all examples simultaneously, or is allowed to cycle through them repeatedly. In many situations, it is
more natural to consider on-line learning paradigms, in which at each time step a new example is chosen.
The examples are never recycled, and the learner is not allowed to simply store them (see e.g, Heskes,
1991; Hansen, 1993; Radons, 1993). Stochastic approximation theory (Kushner, 1978) provides a framework
for understanding of the local convergence properties of on-line learning of smooth functions. This paper
addresses the problem of on-line learning of dichotomies, for which no similarly complete theory yet exists.
304
N. Barkai, H. S. Seung, H. Sompolinsky
We begin with on-line learning of perceptron rules. Since its introduction in the early 60's, the perceptron
algorithm has been used as a simple model of learning a binary classification rule. The algorithm has been
proven to converge in finite time and to yield a half plane separating any set of linearly separable examples.
The perceptron algorithm, however, is not efficient in the sense of distribution-free PAC learning (Valiant,
1984), for one can construct input distributions that require an arbitrarily long convergence time. In a recent
paper (Baum, 1990) Baum proved that the perceptron algorithm applied in an on-line mode, converges as
p-I/3 when learning a half space under a uniform input distribution, where P is the number of presented
examples drawn at random. For on-line learning P is also the number of time steps. Baum also generalized
his result to any "non-malicious" distribution. Kabashima has found the same power law for learning a
two-layer parity machine with non-overlapping inputs, using an on-line least action algorithm (Kabashima,
1994).
If efficiency is measured only by the number of examples used (disregarding time), these particular on-line
algorithms are much worse than batch algorithms. Any batch algorithm which is able to correctly classify a
given set of P examples will converge as p- I (Vapnik, 1982; Amari, 1992; Seung, 1992) . In this paper, we
construct on-line algorithms that can actually achieve the same power law as batch algorithms, demonstrating
that the results of Baum and Kabashima do not reflect a fundamental limitation of on-line learning.
In Section 3, we study on-line algorithms for perceptron learning of a target rule that is not realizable by
a perceptron. Here it is nontrivial to construct an algorithm that even converges to the optimal one, let
alone to optimize the rate of convergence. For the special case of a target rule that is a percept ron corrupted
by output noise this can be done. In Section 4, our results are generalized to dichotomies generated by
thresholding smooth functions. In Section 5 we summarize the results.
2
On-line learning of a perceptron rule
We consider a half space rule generated by a normalized teacher perceptron Wo ERN, Wo' Wo = 1 such
that any vector 5 E RN is given a label uo(5) = sgn(Wo ?5). We study the case of a Gaussian input
distribution centered at zero with a unit variance in each direction in space:
P(5) =
1
s , /2
II _eN
;=1
2
v'21r
(1)
Averages over this input distribution will be written with angle brackets (). A student percept ron W is
trained by an on-line perceptron algorithm. At each time step, an input 5 E RN is drawn at random,
according to distribution Eq. (1) and the student's output u(5) = sgn(W . 5) is calculated. The student is
then updated according to the perceptron rule:
W' = W
+ ~f(5;W)uo(5)5
(2)
and is then normalized so that W . W = 1 at all times. The factor f(5; W) denotes the error of the student
perceptron on the input 5: f = 1 if u(5)uo(5) = 1, and 0 otherwise. The learning rate 1/ is the magnitude
of change of the weights at each time step. It is scaled by N to ensure that the change in the overlap
R = W? Wo is of order lIN. Thus, a change of 0(1) occurs only after presentation of P = O(N) examples.
The performance of the student is measured by the generalization error, defined as the probability of disagreement between the student and the teacher on an arbitrary input fg = (f(5; W?). In the present case,
fg is
cos- I R
(3)
fg = - - 7 r - '
Although for simplicity we analyze below the performance of the perceptron rule (2) only for large N,
our results apply to finite N as well. Multiplying Eq. (2) by Wo after incorporation of the normalization
operation and averaging with respect to the input distribution (1), yields the following differential equation
for R(a) where a = PIN,
(4)
On-line Learning of Dichotomies
Here terms of order
J17/N
305
have been neglected.
The evolution of the overlap R, and thus of the generalization error, depends on the schedule at which the
learning rate 17 decreases. We consider two cases, a constant 17 and a time-dependent 17.
Constant learning rate: When 17 is held fixed, Eq. (4) has a stable fixed point at R < 1, and hence eg
converges to an '7-dependent nonzero value eoo ('7). For '7 ?: 1, 1- Roo ('7) oc '7 2 and eg oc VI - R is therefore
proportional to '7,
(5)
eoo ('7) = '7/..f2i3 .
The convergence to this value is exponential in 0, eg(o) - eoo ('7) '" exp(-'7o/$).
Time-dependent learning rate: Convergence to eg = 0 can be achieved if '7 decreases slowly enough with
o. We study the limiting behaviour of the system for '7 which is decreasing with time as '7 = (7]0$) o-z.
z > 1. In this case the rate is reduced too fast before a sufficient number of examples have been seen. This
results in R which does not converge to 1 but instead to a smaller value that depends on its initial value.
z < 1. The system follows the change in '7 adiabatically. Hence, to first order in
Thus, eg converges to zero with an asymptotic rate eg(o) '" o-z.
z
= 1.
0- 1 ,
eg(o)
= eoo ('7(o?.
The behaviour of the system depends on the prefactor '70:
(1 '7~)1
eg
'70> 1
71"'70- 1 0
y'logo
7]0 = 1
a
A
7]0
0'1<>
(6)
<1
where A depends on the initial condition. Thus the optimal asymptotic change of '7 is 2..;21i/ 0, in which case
the error will behave asymptotically as eg(o) '" 1.27/0. This is not far from the batch asymptotic (Seung,
1992) eg(o) '" 0.625/0. We have confirmed these results by numerical simulation of the algorithm Eq. (2).
Figure 1 presents the results of the optimalleaming schedule, i.e., '7 = 2..;21i/o. The numerical results are
in excellent agreement with the prediction eg(o) = 1.27/0 for the asymptotic behavior. Finally, we note
that our analysis of the time-dependent case is similar to that of Kabashima and Shinomoto for a different
on-line learning problem (Kabashima, 1993).
3
On-line learning of a percept ron with output noise
In the case discussed above, the task can be fully realized by a perceptron, i.e., there is a perceptron W
such that eg = O. In more realistic situations a percept ron will only provide an approximation of the target
function, so that the minimal value of eg is greater than zero. These cases are called unrealizable tasks. A
drawback of the above on-line algorithm is that, for a general unrealizable task, it does not converge to
the optimal perceptron, i.e., it does not approach the minimum of ego To illustrate this fact we consider a
perceptron rule corrupted by output noise. The label of an input S is O"o(S), where O"o(S) = sgn(Wo . S)
with probability 1 - p, and - sgn(Wo . S) with probability p. We assume 0 $ p $ 1/2. For reasons which
will become clear later, the input distribution is taken as a Gaussian centered at U
P(S)
=
n..;21i
N
_1_ e -(S,-u.)'/2
(7)
;=1
In this case eg is given by
eg
=p+ (l-
q
-00
DYH(~)+l?O
DYH(~?).
1- R
1- R
(8)
_q
where qo = U . W 0 denotes the overlap between the center of the distribution and the teacher perceptron,
and q = U . W is the overlap between the center of the distribution and W. The integrals in Eq. (8) are
306
N. Barkai, H. S. Seung, H. Sompolinsky
with respect to a Gaussian measure Dy = exp(-y2/2)/.J21i and H(x) = J.,oo Dy. Note that the optimal
perceptron is the teacher W = Wo i.e., R = 1, q = 110, which yields the minimal error fmin = p.
First, we consider training with the normalized perceptron rule (2). In this case, we obtain differential
equations for two variables: R and q. Solving these equations we find that in general, W converges to a
vector with a direction which is in the plane of Wo and U and is does not point in the direction of Wo even
in the limit of,., -+ O. Here we present the result for the limit of,., -+ 0 and small noise level, i.e., p ?: 1. In
this case, we obtain for f oo ('" = 0)
foo(O)
= P + p (1 - 2H(qo?~ + O(p2)
1+(u2-q~)
(9)
where u = lUI is the magnitude of the center of the input distribution. For p = 0, the only solution is
R = 1 and q = qo, in agreement with the previous results. For p > 0 the optimal solution is retrieved only
in the following special cases: (i) the input distribution is isotropic, i.e., qo = u = OJ (ii) when U is parallel
to W o, i.e., u = qoj and (iii) when U is orthogonal to W o, i.e., qo = O. This holds also for large value of
p . In these special cases, the symmetry of the input distribution relative to the teacher vector, guarantees
that the deviations from W = Wo incurred by the inputs that come with the wrong label cancel each other
on average. According to Eq. (9), for other directions of U, fg is above the optimal value. Note that the
additional term in fg is of the same order of magnitude (O(P? as the minimal error.
In the following we suggest a modified on-line algorithm for learning a perceptron rule with output noise.
The student weights are changed according to
W' = W
+ ~f(S; W)ITo(S)(S -
T(S?
(10)
followed by a normalization of W. This algorithm differs from the perceptron algorithm in that the change in
W is not proportional to the present input, but to a shifted vector. The shifting vector T(S), is determined
by the requirement that the teacher Wo will be a fixed point of the algorithm in the limit of,., -+ O. This is
equivalent to the condition
(11)
(fo(S)lTo(S)(S - T(S?) = 0
where foeS) is the error function for S when W = Wo o This condition does not determine T uniquely. A
simple choice is one for which T is independent of S. This leads to
T
= (sgn(Wo' S)S) = (IToS)
(sgn(Wo . S?
(ITo)
(12)
where we used the fact that for any S, fo(S)lTo(S) equals - sgn(Wo . S) with probability p, and zero with
probability (1 - p). This uniform shift is possible only when (O'o) ~ 0, namely when the average frequencies
of +1 and -1 labels are not equal. If this is not the case, one has to choose nonuniform forms of T(S) .
Note that in general T has to be learned so that Eq. (10) has to be supplemented by appropriate equations
for changing T. In the case of Eq. (12), one can easily learn separately the numerator and denominator by
running averages of O'oS and 0'0, respectively. We have studied analytically the above algorithm for the case
of the Gaussian input distribution Eq. (7), in the limit of large N. The shifting vector is given by
T
= U + Wo ~ exp( -q~/2)
V1T 1 -
2H(qo)
The differential equations for the overlaps R and q in the neighborhood of the point R = 1 and q
daR
da
doq
da
(13)
= qo are,
(14)
where oR = 1 - R and oq = qo - q. In the limit,., -+ 0, R = 1 and q = qo is indeed a stable fixed point of
the algorithm, so that the student converges to the optimal perceptron W o, and hence the feneralization
error converges to its minimal value fmin = p. Since, unlike Eq. (4), the coefficient of the,., term in Eq.
On-line Learning of Dichotomies
307
(14) is constant, c5Roo(1]) <X 1], for small fixed 1], and not to 1]2. Thus, in this case, the generalization error
approaches, in the limit a -t 00, the value
exp(-qa/ 4)
t:oo (1]) :;;: p +.,fiiP (271"3)1/4
(15)
For a time-dependent 1], the convergence to the optimal weights depends on the choice of 1](n), as in the
case of the noiseless perceptron rule. For 1] :;;: (1]0
exp( q5 12) ) a-' , with z ~ 1, the error converges to
p. For z < 1, to first order in 1/a, t:g(a) :;;: t: oo (1](a)) , yielding
J1i72
t:g(a) - p ~ a-?/ 2 ?
(16)
When z :;;: 1, the rate of convergence depends on the value of 1]0.
t:g
and logarithmic corrections to a- 1 / 2 for
( )
a
a-l /2
1] > 1
p ~ { a-'1oj:~, ~ < 1
-
1]0 :;;:
1. Thus, the optimal rate of convergence is
t: (a) _ p
9
which is achieved for
1]0
(17)
~ V~
f2P
(18)
= 2.
We have tested successfully this algorithm by simulations of learning a perceptron rule with output noise
with several input distributions, including the Gaussian, of Eq. (7). Figure 2 presents the generalization
error as a function of a for the Gaussian distribution, with p = 0.2 , and we have chosen 1]0 = 2. The error
converges to the optimal value 0.2 as a- I / 2 in agreement with the theory. For comparison the result of the
usual perceptron algorithm is also presented. This algorithm converges to t:g ~ 0.32, clearly larger than the
optimal value.
4
On-line learning of thresholded smooth functions
Our results for the realizable perceptron can be extended to a more general class of dichotomies, namely
thresholded smooth functions. They are defined as dichotomies of the form
(19)
O"(S; W) = sgn(f(Sj W))
where 1 is a differentiable function of a set of parameters, denoted by W, and S is the input vector. We
consider here the case of a realizable task, where the examples are given with labels 0"0 corresponding to a
target machine W 0 which is in the W space. For this task we propose the following generalization of the
perceptron rule (2)
(20)
W' = W + 1]t:(Sj W)O"o(S)V I(S; W)
where V denotes a gradient w.r.t. W. Then, as we argue below, the vector Wo is a stable fixed point in the
limit of 1] -t O. Furthermore, for constant small 1] the residual error scales as t:oo <X 1]. For 1] ~ a-' , z < 1,
t:g(a) ~ t:oo(1](a)) '""' a-?.
To show this, let us consider for simplicity the one-dimensional case, w' = w + 1]g(w, s), where
81
g(w,s) = 9(-I(w,s)l(wo,s))sgn(f(wo,s))8w .
(21)
This equation can be converted into a Markov equation for the probability distribution, P(w,n) (Van Kampen, 1981)
P(w,n
+ 1):;;:
J
dw'W(w'lw)P(w' , n)
(22)
where W(wlw') =< c5(w' - w -1]g(w, s)) > is the transition rate from w to w'. In the limit of small fixed 1],
the equilibrium distribution, Poo, can be shown to have the following scaling form,
1
Poo(w;1]):;;: -F(c5wl1])
(23)
1]
308
N. Barkai, H. S. Seung, H. Sompolinsky
where ow = w - Wo and F(x) obeys the following difference equation
LF(x) ==
L
9?(f~ + O"x)J~)I(f~ + O"x)IF(x + o"/~) -lxIF(x)
=0
(24)
C1=?1
where J6 is the value of the gradient aJ(wo, s)law at the decision boundary of J(wo, s), namely at the point
s obeying J(wo,s) = O. Note that since we are interested in normalizable solutions of Eq. (24), F(x) has to
vanish for for all x > 1161. This result is valid provided the input distribution is smooth and nonvanishing
near the decision boundary. Furthermore, aJ law at Wo may not vanish on the decision boundary. Under
the same conditions, it can be shown that the error is homogeneous in ow with degree 1, hence it should
scale linearly with "I, i.e., 1':00 oc "I. It should be noted that, unlike other on-line learning problems (Heskes,
1991; Hansen, 1993; Radons, 1993), the equilibrium distribution is our case is not Gaussian.
For a time-dependent "I of the form "I = TJon- z , z < I, pew, n) at long times is of the form
(25)
where F is the stationary distribution, given by Eq. (24) and the coefficient of the correction, G, solves the
inhomogeneous equation
dF
?
(26)
zx dx + zF(x) = TJoLG(x)
where the linear operator L is defined in Eq. (24). Thus, to leading order in inverse time, the system
follows adiabatically the finite-TJ stationary distribution, yielding I':g(n) which vanishes asymptotically as
I':g(n) oc TJ(n) ~ n- z . The optimal schedule is obtained for z = 1. In this case, P(w,n) = "1-1 (n)F(ow/TJ(n))
where F(x) solves the homogeneous equation
dF
zx dx
?
+ zF(x) = TJoLF(x)
For sufficiently large "10, this equation has a solution, implying that
(27)
I':g
oc n- 1
Similarly, the results of Section 3 can also be extended to the case of thresholded- smooth functions with
a probability p of an error due to isotropic output noise. In this case, the optimal choice is again "I oc n- I
yielding I':g -p Rl ,;ri. It should be noted that for this case, the probability distribution for small "I does reduce
to a Gaussian distribution in owl,;ri. Using a multidimensional Markov equation, it is straightforward to
extend these results to higher dimensions. The small "I limit yields equations similar to Eqs. (24-26), that
involve integration over the decision boundary of J(W, S).
5
Summary and Discussion
We have found that the perceptron rule (2) with normalization can lead to a variety of learning curves,
depending on the schedule at which the learning rate is decreased. The optimal schedule leads to an inverse
power law learning curve, I':g ~ 0- 1 . Baum's results (Baum, 1990) of a non-normalized perceptron with a
constant learning rate can be viewed as a special case of the above analysis. In the non-normalized perceptron
algorithm, the magnitude of the student's weights grow with 0 as IWI ~ 0 1 / 3 . The time evolution of the
overlap R, and thus of the generalization error is governed by the effective learning rate TJe/f = T//IWlleading
via Eq. (6) to the result I':g ~ 0- 1 / 3 . Similar results apply to the two-layer parity machine studied in
(Kabashima, 1994).
Our analysis, leading to the equations of motion (4) and (14), was based on the limit of large N and P, such
that 0 = PIN remains finite. We would like to stress however, that this limit is only necessary in deriving
the full form of the learning curve, i.e., R(o) for all o. On the other hand, our results for the large P
asymptote of the learning curve for smailT/ are valid for finite N as well, as implied by the general treatment
of the previous section.
Unrealizable percept ron rules present a more complicated problem. We have presented here a modified
perceptron algorithm that converges to the optimal solution in the special case of an isotropic output noise.
On-line Learning of Dichotomies
309
In this case, the convergence to the optimal error is as 01- 1 / 2 . This is the same power law as obtained in
the standard sample complexity upper bounds (Vapnik, 1982) and in the approximate replica symmetric
calculations (Seung, 1992) for batch learning of unrealizable rules. It should be stressed however, that the
success of the modified algorithm in the case of an output noise depends on the fact that the errors made
by the optimal solution are uncorrelated with the input. Thus, finding an on-line algorithm that can cope
with other types of unrealizability remains an important problem.
The learning algorithms for the perceptron rule, without and with output noise, can be generalized to learning
thresholded smooth functions, assuming certain reasonable properties of the input distribution are present,
as shown in Section 4. The dependence of the learning curve on the learning rate schedule remains roughly
the same as in the percept ron case. This implies that on-line learning of realizable dichotomies, with possible
output noise, can achieve the same power laws in the number of examples that is typical of batch learning
of such rules. Furthermore, the on-line formulation possesses the theoretical virtues of addressing time as
well as sample complexity, so that the same power laws imply the polynomial relationship between the time
and the achieved error level. The above conclusions assume that the equilibrium state at small learning
rates is unique, which in general is not the case. The issue of overcoming local minima in on-line learning
is a difficult problem (Heskes, 1992) Finally, the theoretical results for on-line learning has the important
advantage of not requiring the use of the often problematic replica formalism.
Acknowledgements
We are grateful for helpful discussions with Y. Freund, M. Kearns, R. Schapire, and E. Shamir, and thank Y.
Kabashima for bringing his paper to our attention. HS is partially supported by the Fund for Basic Research
of the Israeli Academy of Arts and Sciences.
References
S. Amari, N. Fujita, and S. Shinomoto. Four types of learning curves. Neural Comput., 4:605-618, 1992.
E. B. Baum. The perceptron algorithm is fast for nonmalicious distributions. Neural Comput., 2:248-260,
1990.
H. J. Kushner and D. S. Clark. Stochastic approximation methods for constrained and unconstrained systems.
Springer, Berlin, 1978.
L. K. Hansen, R. Pathria, and P. Salamon. Stochastic dynamics of supervised learning. J. Phys., A26:63-71,
1993.
T . Heskes and B. Kappen. Learning processes in neural networks. Phys. Rev., A44:2718-2762, 1991.
T. Heskes, E. T. P. Slijpen, and B. Kappen. Learning in neural networks with local minima. Phys. Rev.,
A46:5221-5231, 1992.
Y. Kabashima. Perfect loss of generalization due to noise in k = 2 parity machines. J. Phys., A27:1917-1927,
1994.
Y. Kabashima and S. Shinomoto. Incremental learning with and without queries in binary choice problems.
In Proc. of IJCNN, 1993.
G. Radons. On stochastic dynamics of supervised learning. J. Phys., A26:3455-3461, 1993.
H. S. Seung, H. Sompolinsky, and N. Tishby. Statistical mechanics of learning from examples. Phys. Rev.,
A45:6056-6091, 1992.
L. G. Valiant. A theory of the learnable. Commun. ACM, 27:1134-1142,1984.
N. G. Van Kampen. Stochastic processes in physics and chemistry. North holland 1981.
V. N. Vapnik. Estimation of Dependences based on Empirical Data. Springer-Verlag, New York, 1982.
N. Barkai, H. S. Seung, H. Sompolinsky
310
0.03
0.025
0.02
0.015
f9
0.01
0.005
0
0.005
0
0.01
0.015
0.02
1/a
Figure 1: Asymptotic performance of a realizable perceptron. Simulation results for 110 :;;: 2 and N :;;: 50
(solid curve) are compared with the theoretical prediction f g :;;: 1.271a (dashed curve).
0.35
0.3
0.25
0.2
fg -
.'
P
.....
....
0.15
o
?
?
'c"..'
..
..
.. ..'' . ' .. .
'
'.:....
~
0.1
0.05
0
0
0.1
0.2
0.3
lIfo
0.4
0.5
0.6
0.7
Figure 2: Simulation results for on-line learning of a perceptron with output noise. Here 1Jo :;;: 2, P :;;: 0.2,
U = 4, and qo :;;: -1.95. The regular percept ron learning (dashed curve) is compared with the
modified algorithm (solid curve). The dashed line shows the theoretical prediction Eq. (18)
N :;;: 250,
| 976 |@word h:1 polynomial:1 simulation:4 solid:2 kappen:2 initial:2 att:1 com:1 yet:1 dx:2 written:1 realistic:1 numerical:2 asymptote:1 fund:1 alone:1 half:3 stationary:2 implying:1 plane:2 isotropic:3 provides:1 ron:7 differential:3 become:1 indeed:1 roughly:1 behavior:1 tje:1 mechanic:1 v1t:1 adiabatically:2 decreasing:1 begin:1 provided:1 lowest:1 israel:2 finding:1 nj:1 guarantee:1 multidimensional:1 scaled:1 wrong:1 unit:1 uo:3 before:1 local:3 limit:11 logo:1 studied:3 co:1 obeys:1 unique:1 differs:1 lf:1 empirical:1 bell:2 regular:1 suggest:1 operator:1 optimize:1 equivalent:2 center:3 baum:7 poo:2 jerusalem:2 straightforward:1 attention:1 focused:1 simplicity:2 rule:20 deriving:1 his:2 dw:1 racah:2 updated:1 limiting:1 target:7 shamir:1 homogeneous:2 agreement:3 ego:1 prefactor:1 cycle:1 sompolinsky:6 decrease:4 vanishes:1 complexity:2 seung:9 neglected:1 dynamic:2 trained:1 q5:1 solving:1 grateful:1 efficiency:1 learner:2 lto:2 easily:1 fast:5 effective:1 query:1 dichotomy:10 neighborhood:1 larger:1 roo:1 amari:2 otherwise:1 advantage:1 differentiable:1 propose:2 achieve:2 wlw:1 academy:1 convergence:9 requirement:1 perfect:1 converges:12 incremental:1 illustrate:1 oo:6 ac:2 depending:1 measured:2 eq:18 solves:2 p2:1 come:1 implies:1 direction:4 inhomogeneous:1 drawback:1 stochastic:5 centered:2 sgn:9 owl:1 require:1 eoo:4 behaviour:2 generalization:10 unrealizable:4 correction:2 hold:1 sufficiently:1 exp:5 equilibrium:3 early:1 estimation:1 proc:1 label:5 hansen:3 successfully:1 clearly:1 gaussian:8 modified:4 lifo:1 nonmalicious:1 realizable:6 sense:1 helpful:1 dependent:6 interested:1 fujita:1 issue:1 classification:1 denoted:1 art:1 special:5 integration:1 constrained:1 equal:2 once:1 never:1 construct:3 cancel:1 simultaneously:1 bracket:1 yielding:4 tj:3 held:1 integral:1 necessary:1 orthogonal:1 theoretical:4 minimal:4 classify:1 formalism:1 deviation:1 addressing:1 uniform:2 too:1 tishby:1 teacher:6 corrupted:2 fundamental:1 huji:2 physic:3 f2p:1 nonvanishing:1 jo:1 again:1 reflect:1 choose:1 slowly:1 worse:1 leading:2 converted:1 chemistry:1 student:9 north:1 coefficient:2 depends:8 vi:1 later:1 a45:1 analyze:1 parallel:1 complicated:1 iwi:1 il:2 variance:1 percept:7 yield:4 generalize:1 multiplying:1 confirmed:1 zx:2 j6:1 kabashima:9 foe:1 fo:2 phys:6 frequency:1 proved:1 treatment:1 schedule:8 actually:1 salamon:1 higher:1 dyh:2 supervised:2 formulation:1 done:1 furthermore:3 hand:1 qo:10 o:1 overlapping:1 mode:1 aj:2 behaved:1 barkai:5 normalized:5 y2:1 requiring:1 evolution:2 hence:4 analytically:1 symmetric:1 laboratory:2 nonzero:1 eg:15 shinomoto:3 a46:1 numerator:1 uniquely:1 noted:2 oc:6 generalized:3 hill:1 stress:1 complete:1 motion:1 rl:1 extend:1 discussed:1 pew:1 unconstrained:1 heskes:5 similarly:2 lowered:1 stable:3 kampen:2 recent:1 retrieved:1 commun:1 store:1 certain:1 verlag:1 binary:2 arbitrarily:1 success:1 seen:1 minimum:3 greater:1 additional:1 converge:5 paradigm:1 determine:1 dashed:3 ii:2 full:1 smooth:8 calculation:1 long:2 lin:1 prediction:3 basic:1 denominator:1 noiseless:1 df:2 normalization:3 achieved:3 c1:1 separately:1 decreased:1 grow:1 malicious:1 a44:1 unlike:2 posse:1 bringing:1 oq:1 near:1 iii:1 enough:1 variety:1 reduce:1 shift:1 wo:26 york:1 repeatedly:1 action:1 dar:1 generally:1 clear:1 involve:1 reduced:1 schapire:1 problematic:1 shifted:1 correctly:1 four:1 demonstrating:1 drawn:2 changing:1 thresholded:5 replica:2 asymptotically:2 angle:1 inverse:2 reasonable:1 decision:4 dy:2 scaling:1 radon:3 layer:2 bound:1 followed:1 nontrivial:1 ijcnn:1 incorporation:1 normalizable:1 ri:2 fmin:2 separable:1 ern:1 according:4 smaller:1 rev:3 taken:1 equation:14 remains:3 pin:2 operation:1 apply:2 appropriate:1 disagreement:1 batch:7 denotes:3 running:1 kushner:2 ensure:1 a26:2 murray:1 implied:1 realized:1 occurs:1 dependence:2 usual:1 gradient:2 ow:3 thank:1 separating:1 berlin:1 argue:1 reason:1 assuming:1 relationship:1 hebrew:2 difficult:1 zf:2 upper:1 markov:2 finite:5 behave:1 situation:2 extended:2 rn:2 nonuniform:1 arbitrary:1 overcoming:1 namely:3 optimized:1 learned:1 qa:1 address:1 able:1 israeli:1 below:2 summarize:1 oj:2 including:1 shifting:2 power:6 overlap:6 natural:1 residual:1 imply:1 understanding:1 acknowledgement:1 asymptotic:5 law:8 relative:1 fully:1 freund:1 loss:1 limitation:1 proportional:2 proven:1 clark:1 incurred:1 degree:1 sufficient:1 thresholding:1 uncorrelated:1 changed:1 summary:1 supported:1 parity:3 free:1 perceptron:41 institute:2 fg:6 van:2 curve:14 calculated:1 boundary:4 transition:1 valid:2 dimension:1 c5:1 made:1 far:1 cope:1 sj:2 approximate:1 learn:1 recycled:1 symmetry:1 excellent:1 da:2 linearly:2 noise:14 allowed:2 f9:1 en:1 foo:1 obeying:1 exponential:1 comput:2 governed:1 vanish:2 lw:1 ito:2 pac:1 supplemented:1 learnable:1 disregarding:1 virtue:1 exists:1 vapnik:3 valiant:2 magnitude:4 logarithmic:1 simply:1 partially:1 u2:1 holland:1 springer:2 acm:1 viewed:1 presentation:1 change:6 determined:1 typical:1 lui:1 averaging:1 kearns:1 called:1 stressed:1 tested:1 |
7,217 | 977 | New Algorithms for
2D and 3D Point Matching:
Pose Estimation and Correspondence
Steven Gold l , Chien Ping LuI, Anand Rangarajan l ,
Suguna Pappu l and Eric Mjolsness 2
Department of Computer Science
Yale University
New Haven, CT 06520-8285
Abstract
A fundamental open problem in computer vision-determining
pose and correspondence between two sets of points in spaceis solved with a novel, robust and easily implementable algorithm .
The technique works on noisy point sets that may be of unequal
sizes and may differ by non-rigid transformations. A 2D variation calculates the pose between point sets related by an affine
transformation-translation, rotation, scale and shear. A 3D to 3D
variation calculates translation and rotation . An objective describing the problem is derived from Mean field theory. The objective
is minimized with clocked (EM-like) dynamics. Experiments with
both handwritten and synthetic data provide empirical evidence
for the method.
1
Introduction
Matching the representations of two images has long been the focus of much research
in Computer Vision, forming an essential component of many machine-based ob1 E-mail address of authors: lastname-firstname@cs.yale.edu
2Department of Computer Science and Engineering, University of California at San
Diego (UCSD), La Jolla, CA 92093-0114. E-mail: emj@cs.ucsd.edu
958
Steven Gold, Chien Ping Lu, Anand Rangarajan, Suguna Pappu, Eric Mjolsness
ject recognition systems. Critical to most matching techniques is the determination
of correspondence between spatially localized features within each image. This
has traditionally been considered a hard problem - especially when the issues of
noise, missing or spurious data, and non-rigid transformations are tackled [Grimson, 1990] . Many approaches have been tried, with tree-pruning techniques and
generalized Hough transforms being the most common. We introduce anew, robust and easily implementable algorithm to find such poses and correspondences.
The algorithm can determine non-rigid transformations between noisy 2D or 3D
spatially located unlabeled feature sets despite missing or spurious features. It is
derived by minimizing an objective function describing the problem with a combination of optimization techniques, incorporating Mean Field theory, slack variables,
iterative projective scaling, and clocked (EM-like) dynamics.
2
2D with Affine Transformations
2.1
Formulating the Objective
Our first algorithm calculates the pose between noisy, 2D point sets of unequal size
related by an affine transformation - translation, rotation, scale and shear. Given
two sets of points {Xj} and {Yk}, one can minimize the following objective to find
the affine transformation and permutation which best maps Y onto X :
J
E2D(m, t, A) =
K
L L mjkllXj -
J
t - AYkll2 + g(A) -
j=l k=l
K
aLL mjk
j=l k=l
with constraints: Vj Ef=l mjk ~ 1 , Vk Ef=l mjk ~ 1 , Vjk mjk ~ 0 and
g(A)
= 1a2 + /'i,b 2 + AC2
A is decomposed into scale, rotation, vertical shear and oblique shear as follows:
where,
s(a) = ( eOa
o )
e- b
,Sh 2 (c)
cosh(c)
= ( sm
. h( c)
sinh(c) )
cosh(c)
R(8) is the standard 2x2 rotation matrix. g(A) serves to regularize the affine transformation - bounding the scale and shear components. m is a fuzzy correspondence
matrix which matches points in one image with corresponding points in the other
image. The constraints on m ensure that each point in each image corresponds to
at most one point in the other image. However, partial matches are allowed, in
which case the sum of these partial matches may add up to no more than one. The
inequality constraint on m permits a null match or multiple partial matches.
The a term biases the objective towards matches. The decomposition of A in the
above is not required, since A could be left as a 2x2 matrix and solved for directly in
the algorithm that follows . The decomposition just provides for more precise regularization, i.e., specification of the likely kinds oftransformations. Also Sh2(C) could
New Algorithms for 2D and 3D Point Matching
959
be replaced by another rotation matrix, using the singular value decomposition of
A.
We transform the inequality constraints into equality constraints by introducing
slack variables, a standard technique from linear programming;
K
L
Vj
K+1
mj k ::;
1
-+
Vj
k=l
L
mj k
= 1
k=l
and likewise for the column constraints. An extra row and column are added to
the matrix m to hold the slack variables. Following the treatment in [Peterson and
Soderberg, 1989; Yuille and Kosowsky, 1994] we employ Lagrange multipliers and
an x log x barrier function to enforce the constraints with the following objective:
J
E2D(m, t, A)
K
J
=L
L mjkllXj j=l k=l
1 J+1K+1
+~
L:
J
L mik(logmjk -1)
i=l k=l
t - AYk ll
2
K+1
+ g(A) K
K
ct L
L mjk
j=l k=l
J+l
+ LJlj(L mjk -1) + LlIk(L mjk -1)
i=l
k=l
k=l
(1)
j=l
In this objective we are looking for a saddle point. (1) is minimized with respect to
m, t, and A which are the correspondence matrix, translation, and affine transform,
and is maximized with respect to Jl and 1I, the Lagrange multipliers that enforce
the row and column constraints for m. m is fuzzy, with the degree of fuzziness
dependent upon f3.
2.2
The Algorithm
The algorithm to minimize the above objective proceeds in two phases. In phase
one, while {t, A} are held fixed, m is initialized with a coordinate descent step,
described below, and then iteratively normalized across its rows and columns until
the procedure converges (iterative projective scaling). This phase is analogous to a
softmax update, except that instead of enforcing a one-way, winner-take-all (maximum) constraint, a two-way, assignment constraint is being enforced. Therefore
we describe this phase as a softassign. In phase two {t, A} are updated using coordinate descent. Then f3 is increased and the loop repeats. Let E2D be the above
objective (1) without the terms that enforce the constraints (i.e. the x log x barrier
function and the Lagrange parameters).
In phase one (softassign) m is updated via coordinate descent:
mjk
E
2D
= exp (-f3-a
-)
amjk
Then m is iteratively normalized across j and k until Ef=l
Ef=l Llmiajk < i
:
Using coordinate descent the {t, A} are updated in phase two. If a term of {A}
cannot be computed analytically (because of its regularization), Newton's method
960
Steven Gold, Chien Ping Lu, Anand Rangarajan, Suguna Pappu, Eric Mjolsness
is used to compute the root of the function . So if a is a term of {t, A} then in phase
two we update a such that 8!~D = O. Finally f3 is increased and the loop repeats.
By setting the partial derivatives of E2D to zero and initializing the Lagrange parameters to zero, the algorithm for phase one may be derived . Beginning with a
small f3 allows minimization over a fuzzy correspondence matrix m, for which a
global minimum is easier to find. Raising f3 drives the m's closer to 0 or 1, as the
algorithm approaches a saddle point.
3
3D with Rotation and Translation
The second algorithm solves the 3D-3D pose estimation problem with unknown
correspondence. Given two sets of 3D points {Xj} and {Yk} find the rotation R,
translation T, and correspondence m that minimize
K
J
E3D(m ,T,R)
J
K
= LLmjkllRXj +T-YkI12-aLLmjk
j=l k=1
j=l k=1
with the same constraint on the fuzzy correspondence matrix m as in 2D affine
matching. Note that there is no regularization term for the T - R parameters.
This algorithm also works in two phases. In the first, m is updated by a soft assign
as was described for 2D affine matching. In the second phase, m is fixed, and the
problem becomes a 3D to 3D pose estimation problem formulated as a weighted least
squares problem. The rotation and translation are represented by a dual number
quaternion (r, s) which corresponds to a screw coordinate transform [Walker et al. ,
1991] . The rotation can be written as R(r) = W(r)tQ(r) and the translation as
W(r)ts . Using these representations, the objective function becomes
J
K
E3D = L L mjkllW(r)tQ(r)xj
j=1 k=1
+ W(r)t s - Ykl1 2
where Xj = (Xj, 0)1 and Yk = (Yk,O)t are the quaternion representations of Xj
and Yk, respectively. Using the properties that Q(a)b = W(b)a and Q(a)tQ(a) =
W(a)tW(a) = (ata)J, the objective function can be rewritten as
(2)
where
J
K
- L L mjkQ(Yk)tW(Xj)
j=1k=1
1
J
K
2LLmjkI
j=1k=1
J
C3
=
K
L L mjk(W(Xj) - Q(Yk)).
j=1k=1
New Algorithms for 2D and 3D Point Matching
961
With this new representation, all the information, including the current fuzzy estimate of the correspondence m are absorbed into the three 4-by-4 matrices Cl, C2 , C3
in (2), which can be minimized in closed-form [Walker et al., 1991].
4
Experimental Results
In this section we provide experimental results for both the 20 and 30 matching
problems. As an application of the 20 matching algorithm, we present results in
the context of handwritten character recognition.
4.1
Handwritten Character Data
The data were generated using an X-windows tool which enables us to draw an
image with the mouse on a writing pad on the screen. The contours of the images
are discretized and are expressed as a set of points in the plane. In the experiments
below, we generate 70 points per character on average.
The inputs to the point matching algorithm are the x-y coordinates generated by the
drawing program. No other pre-processing is done. The output is a correspondence
matrix and a pose. In Figures 1 and 2, we show the correspondences found between
several images drawn in this fashion.To make the actual point matches easier to
see, we have drawn the correspondences only for every other model point.
~
:..... -_........ ..........?
_
Figure 1: Correspondence of digits
In one experiment, we drew examples of individual digits, one as a model digit
and then many different variations of it. In Figure 1, it can be seen that the
Steven Gold, Chien Ping Lu, Anand Rangarajan, Suguna Pappu, Eric Mjolsness
962
..~
?
??
??
?\.
I
?
?
????
??
??
?????
? ?
.."..: ....
... .. ..??
..,. ..,..
?
??
?
?
' ?? J
?
??
?
Figure 2: Correspondence: "a" found in "cat", "0" found in "song"
correspondences are good for a large variation from the model digit. For example,
the correspondence is invariant to scale. Also, the correspondence is good between
distorted digits, as in 3 and 6, or between different forms of a digit as in 4, 3, and
2.
In another experiment (Figure 2), individual letters are correctly identified within
words. Here, no pre-processing to segment the cursive word into letters is done.
The correspondence returned by the point matching algorithm by itself can be good
enough for identification. Even similar letters may be differentiated, for example
the "a" in "cat" is correctly identified even though the "e" has a similar shape and
the "0" is correctly identified in "song" , despite the similarity of the "s".
4.2
Randomly generated point sets: 2D
In the second set of experiments, randomly generated dot patterns were used. In
each trial a model is created by randomly generating with a uniform distribution, 50
points on a grid of unit area. Independent Gaussian noise N(O, 0-) is added to each of
the points creating a jittered image. Then a fraction, Pd, of points are deleted, and
a fraction, P6, of spurious points are added, randomly on the unit square. Finally,
a randomly generated transformation is applied to the set to generate a new image.
The objective then is to recover the transformation and correspondence between
the transformed image and the original point set.
The transformations we have considered are A -+ (Translation, rotation, scale) and
the full affine transformation, A -+ (Translation, rotation, scale, vertical shear,
oblique shear) The transformation parameters, {tz, t y , (J, a, b, c} are bounded in the
following way: -0.5 < t z , ty < 0.5, -27 0 < (J < 27 0 , 0.5 ~ ea ~ 2 where a is the
scale parameter, and 0.7 ~ eb,ee ~ 1/0.7 where b,c are the parameters for the two
shears. Each of the parameters is chosen independently and uniformly from the
possible ranges.
ocrt.a.1
e ? .,ima"te
?
We use the error measure ea = 31 a
w'id~h",
1where ea IS the error measure
for parameter a and width a is the range of permissible values for a . Dividing by
width a is preferable to dividing by aaetual, which incorrectly weights small aaetual
values. The reported error (y axes of Figure 3) is the average error over all the
parameters.
New Algorithms for 2D and 3D Point Matching
963
The time to recover the correspondence and transformation for a problem instance
of 50 points is about 50 seconds on a Silicon Graphics workstation with a R4400
processor. By varying parameters such as the annealing rate or stopping criterion,
this can be reduced to about 20 seconds with some degradation in accuracy. For
each trial combinations of u E {0.01, 0.02, ... , 0.08} and Pd E {O%, 10%,30%, 50%}
and P6 E {O%, 10%} were used.
Results are reported separately for transformations A and A. For each combination of (U,Pd,P6) 500 test instances were generated. Each data point in Figures
3.a and 3.b represents the average error measure for these 500 experiments. The
noise and/or deletion-addition factor increases the error measure monotonically. As
expected, the transformation A has better results than the affine transformation A.
Translation, rotation, scale
Affine
0.6r---~----~----~----~
~
JE
JE
~
~
e
JE
JE
IE
JE
JE
?
+
+
0.2
~
w
? ? ?
~
oS
Q)
E 0.4
eas
v
0.6r---~----~----~----~
o
o
~
+ + + +
g
~
0.02
~
0.04
+ + )( )(
~
~
0.06
oS
Q)
E 0.4
eas
~
co
.. 0.2
?
JE
JE
+
+ + +
~
w
o
o
0.08
Standard deviation of noise
IE
+ +
~
e
?
QS()()(
0.02
0.04
0.06
0.08
Standard deviation of noise
Figure 3: 2D Results for Synthetic Data
x: Pd
= 0.0,P6 = 0.0,
+ : Pd = 0.3,P6 = 0.1,
4.3
0:
Pd
= O.l,p& = 0.1
* : Pd = 0.5,p& =
0.1
Randomly generated point sets: 3D
A test instance for 3D point matching involves generating a random 3D point set
as a model image, and then generating a test image by applying a random transformation, adding noise and then randomly deleting points.
20 points are generated uniformly within an unit cube. The parameters for the
transformation are generated as follows: The three rotation angles for R are selected
from a uniform distribution U[20, 70]. Translation parameters T~, Ty, Tz are selected
from a uniform distribution U[2.5,7.5]. Gaussian noise N(O, u) is added to the
points. The objective then is to recover the three translation and three rotation
parameters and to find the correspondence between this and the original point set.
The results are summarized in Figure 4.
964
Steven Gold, Chien Ping Lu, Anand Rangarajan, Suguna Pappu, Eric Mjolsness
30
... 0.8
X
...
0
t:: 20
W
C
0
X
0
'.0:;
~ 10
a:
~
X
X
0
0
~
~
+
-+
0
0
0.2
......
WO.6
X
0
~
+
0.4
0
X
X
0
?.0:;
J2
en
c
as
~
0
0
+
00
0.6
0
0
0
0.4
.= 0.2
X
X
c
0
X
~
~
+
~
+
0.2
~
+
+
0.4
~
+
0.6
Standard deviation of noise
Standard deviation of noise
Figure 4: 3D Results for Synthetic Data
x: Pd = 0.0,P8 = 0.0,
0.2,P8 = 0.2,
+ : Pd =
5
Pd = 0.1,P8 = 0.1
*: Pd = 0.3, ps = 0.3
0 :
Conclusion
We have developed an algorithm for solving 2D and 3D correspondence problems.
The algorithm handles significant noise, missing or spurious features , and nonrigid transformations. Moreover it works with point feature data alone; inclusion
of other types of feature information could improve its accuracy and speed . This
approach may also be extended to solve multi-level problems. Additionally, the
affine transform might be modified to include higher order transformations. It may
also be used as a distance measure in learning [Gold et al.,1994] .
Acknowledgements
This work has been supported by AFOSR grant F49620-92-J-0465, ONR/DARPA
grant N00014-92-J-4048 , and the Yale Center for Theoretical and Applied Neuroscience (CTAN) . Jing Yan developed the handwriting interface.
References
S. Gold , E. Mjolsness and A. Rangarajan. (1994) Clustering with a domain-specific
distance measure. In J.D. Cowan et al., (eds.), NIPS 6. Morgan Kaufmann.
E. Grimson, (1990) Object Recognition by Computer, Cambridge, MA : MIT Press
C. Peterson and B. Soderberg. (1989) A new method for mapping optimization
problems onto neural networks, Int. Journ . of Neural Sys., 1(1) :3:22.
M. W. Walker, L. Shoo and R. Volz . (1991) Estimating 3-D location parameters
using dual number quaternions, CVGIP: Image Understanding 54(3) :358-367.
A. L. Yuille and J. J . Kosowsky. (1994) . Statistical physics algorithms that converge. Neural Computation, 6:341-356.
| 977 |@word trial:2 open:1 tried:1 decomposition:3 current:1 written:1 shape:1 enables:1 update:2 alone:1 selected:2 plane:1 beginning:1 sys:1 oblique:2 provides:1 location:1 c2:1 vjk:1 introduce:1 p8:3 expected:1 multi:1 ctan:1 discretized:1 decomposed:1 actual:1 window:1 becomes:2 estimating:1 bounded:1 moreover:1 null:1 kind:1 fuzzy:5 developed:2 transformation:21 every:1 preferable:1 unit:3 grant:2 engineering:1 despite:2 id:1 might:1 eb:1 co:1 projective:2 range:2 pappu:5 mjkllxj:2 digit:6 procedure:1 area:1 empirical:1 yan:1 matching:13 pre:2 word:2 onto:2 unlabeled:1 cannot:1 context:1 applying:1 writing:1 map:1 missing:3 center:1 independently:1 q:1 regularize:1 handle:1 variation:4 traditionally:1 coordinate:6 analogous:1 updated:4 diego:1 programming:1 recognition:3 located:1 steven:5 solved:2 initializing:1 mjolsness:6 yk:7 grimson:2 pd:11 dynamic:2 solving:1 segment:1 yuille:2 upon:1 eric:5 easily:2 darpa:1 represented:1 cat:2 describe:1 solve:1 drawing:1 transform:4 noisy:3 itself:1 j2:1 loop:2 gold:7 rangarajan:6 p:1 jing:1 generating:3 converges:1 object:1 pose:8 solves:1 dividing:2 c:2 involves:1 differ:1 assign:1 hold:1 considered:2 exp:1 mapping:1 a2:1 estimation:3 tool:1 weighted:1 minimization:1 mit:1 gaussian:2 modified:1 varying:1 derived:3 focus:1 ax:1 vk:1 dependent:1 rigid:3 stopping:1 pad:1 spurious:4 journ:1 transformed:1 issue:1 dual:2 softmax:1 cube:1 field:2 f3:6 represents:1 emj:1 minimized:3 screw:1 haven:1 ac2:1 employ:1 randomly:7 individual:2 ima:1 replaced:1 phase:11 softassign:2 tq:3 sh:1 held:1 closer:1 partial:4 tree:1 hough:1 initialized:1 logmjk:1 theoretical:1 increased:2 column:4 soft:1 instance:3 assignment:1 introducing:1 deviation:4 uniform:3 graphic:1 reported:2 jittered:1 synthetic:3 fundamental:1 ie:2 physic:1 mouse:1 tz:2 creating:1 derivative:1 summarized:1 int:1 root:1 closed:1 recover:3 minimize:3 square:2 accuracy:2 kaufmann:1 likewise:1 maximized:1 handwritten:3 identification:1 lu:4 drive:1 r4400:1 processor:1 ping:5 ed:1 ty:2 workstation:1 handwriting:1 treatment:1 ea:5 higher:1 done:2 though:1 just:1 p6:5 until:2 o:2 suguna:5 normalized:2 multiplier:2 regularization:3 equality:1 analytically:1 spatially:2 iteratively:2 ll:1 width:2 lastname:1 clocked:2 criterion:1 generalized:1 nonrigid:1 interface:1 image:15 novel:1 ef:4 common:1 rotation:15 shear:8 sh2:1 winner:1 jl:1 silicon:1 significant:1 cambridge:1 grid:1 inclusion:1 dot:1 specification:1 similarity:1 add:1 jolla:1 n00014:1 inequality:2 onr:1 seen:1 minimum:1 morgan:1 converge:1 determine:1 monotonically:1 multiple:1 full:1 match:7 determination:1 long:1 calculates:3 vision:2 addition:1 separately:1 annealing:1 singular:1 walker:3 permissible:1 extra:1 cowan:1 anand:5 ee:1 enough:1 xj:8 identified:3 ject:1 wo:1 song:2 returned:1 cursive:1 transforms:1 cosh:2 reduced:1 generate:2 neuroscience:1 per:1 correctly:3 drawn:2 deleted:1 fraction:2 sum:1 enforced:1 angle:1 letter:3 distorted:1 draw:1 scaling:2 ct:2 sinh:1 tackled:1 yale:3 correspondence:24 constraint:12 x2:2 speed:1 formulating:1 department:2 combination:3 across:2 em:2 character:3 tw:2 invariant:1 describing:2 slack:3 serf:1 rewritten:1 permit:1 enforce:3 differentiated:1 original:2 clustering:1 ensure:1 include:1 newton:1 cvgip:1 especially:1 objective:14 added:4 distance:2 mail:2 enforcing:1 minimizing:1 unknown:1 vertical:2 sm:1 implementable:2 descent:4 t:1 incorrectly:1 extended:1 looking:1 precise:1 ucsd:2 required:1 kosowsky:2 c3:2 raising:1 california:1 unequal:2 deletion:1 nip:1 address:1 proceeds:1 below:2 pattern:1 firstname:1 program:1 including:1 deleting:1 critical:1 mjk:9 improve:1 created:1 understanding:1 acknowledgement:1 determining:1 afosr:1 permutation:1 localized:1 soderberg:2 degree:1 affine:12 translation:13 row:3 ata:1 repeat:2 supported:1 bias:1 mik:1 peterson:2 barrier:2 f49620:1 contour:1 author:1 ayk:1 san:1 pruning:1 chien:5 anew:1 global:1 iterative:2 additionally:1 mj:2 robust:2 ca:1 cl:1 domain:1 vj:3 bounding:1 noise:10 allowed:1 je:8 en:1 screen:1 fashion:1 specific:1 evidence:1 essential:1 incorporating:1 adding:1 drew:1 te:1 easier:2 likely:1 saddle:2 forming:1 absorbed:1 lagrange:4 expressed:1 corresponds:2 ma:1 formulated:1 fuzziness:1 towards:1 hard:1 except:1 lui:1 uniformly:2 degradation:1 experimental:2 la:1 quaternion:3 |
7,218 | 978 | Pattern Playback in the '90s
Malcolm Slaney
Interval Research Corporation
180 l-C Page Mill Road,
Palo Alto, CA 94304
malcolm@interval.com
Abstract
Deciding the appropriate representation to use for modeling human
auditory processing is a critical issue in auditory science. While engineers have successfully performed many single-speaker tasks with LPC
and spectrogram methods, more difficult problems will need a richer
representation. This paper describes a powerful auditory representation
known as the correlogram and shows how this non-linear representation
can be converted back into sound, with no loss of perceptually important information. The correlogram is interesting because it is a neurophysiologically plausible representation of sound. This paper shows
improved methods for spectrogram inversion (conventional pattern
playback), inversion of a cochlear model, and inversion of the correlogram representation.
1 INTRODUCTIONl
My interest in auditory models and perceptual displays [2] is motivated by the problem of
sound understanding, especially the separation of speech from noisy backgrounds and
interfering speakers. The correlogram and related representations are a pattern space
within which sounds can be "understood" and "separated" [3][4]. I am therefore interested
in resynthesizing sounds from these representations as a way to test and evaluate sound
separation algorithms, and as a way to apply sound separation to problems such as speech
enhancement. The conversion of sound to a correlogram involves the intermediate representation of a cochleagram, as shown in Figure 1, so cochlear-model inversion is
addressed as one piece of the overall problem.
1. Much of this work was performed by Malcolm Slaney, Daniel Naar and Richard F. Lyon while all three were employed at Apple Computer. The mathematical
details of this work were presented at the 1994ICASSP[I].
828
Malcolm Slaney
Waveform
Cochleagram
j 1r- *T'T~" "I-~!"'T!-~ -'-o I-'1 ::::S:ec::ti:n:: h
------------~
Correlogram
. . t--'~~'--t.:lfga..~l!!!Jnme
------------~
Figure 1. Three stages in low-level auditory perception are shown here. Sound waves are converted into a detailed representation with broad spectral bands, known as cochleagrams. The
correlogram then summarizes the periodicities in the cochleagram with short-time autocorrelation. The result is a perceptual movie synchronized to the acoustic signal. The two inversion
problems addressed in this work are indicated with arrows from right to left
There are three factors which can be used to judge the quality of an auditory model: psychoacoustic comparisons, neurophysiological plausibility, and does it represent the perceptually relevant information? First, the correlogram has been shown to simply and
accurately predict human pitch perception [5]. The neurophysiological basis for the correlogram has not been found, but there are neural circuits performing the same calculation in
the mustached bat's echolocation system [6]. Finally, from an information representation
point of view, does the correlogram preserve the salient information? The results of this
paper show that no information has been lost. Since the psychoacoustic, neurophysiological, and information representation measures are all positive, perhaps the correlogram is
the basis of most auditory processing.
The inversion techniques described here are important because they allow us to readily
evaluate the results of sound separation models that "zero out" unwanted portions of the
signal in the correlogram domain. This work extends the convex projection approach of
Irino [7] and Yang [8] by considering a different cochlear model, and by including the correlogram inversion. The convex projection approach is well suited to "filling in" missing
information. While this paper only describes the process for one particular auditory
model, the techniques are equally useful for other models.
This paper describes three aspects of the problem: cochleagram inversion, conversion of
the correlogram into spectrograms, and spectrogram inversion. A number of reconstruction options are explored in this paper. Some are fast, while other techniques use time-consuming iterations to produce reconstructions perceptually equivalent to the original sound.
Fast versions of these algorithms could allow us to separate a speaker's voice from the
background noise in real time.
2 COCHLEAGRAM INVERSION
Figure 2 shows a block diagram of the cochlear model [9] that is used in this work. The
basis of the model is a bank of filters, implemented as a cascade of low-pass filters, that
splits the input signal into broad spectral bands. The output from each filter in the bank is
called a channel. The energy in each channel is detected and used to adjust the channel
gain, implementing a simple model of auditory sensitivity adaptation, or automatic gain
control (AGe). The half-wave rectifier (HWR) detection nonlinearity provides a waveform for each channel that roughly represents the instantaneous neural firing rate at each
position along the cochlea.
E
~---+
>
~
Cochlear
Filterbank
Detector
(HWR)
Adaptation
or
AGC
Figure 2. Three stages of the simple cochlear model used in this paper are shown above.
Pattern Playback in the '90s
829
The cochleagram is converted back into sound by reversing the three steps shown in Figure 2. First the AGe is divided out, then the negative portions of each cochlear channel are
recovered by using the fact that each channel is spectrally limited. Finally, the cochlear filters are inverted by running the filters backwards, and then correcting the resulting spectral slope.
The AGe stage in this cochlear model is controlled by its own output. It is a combination
of a multiplicative gain and a simple first-order filter to track the history of the output signal. Since the controlling signal is directly available, the AGe can be inverted by tracking
the output history and then dividing instead of multiplying. The performance of this algorithm is described by Naar [10] and will not be addressed here. It is worth noting that AGe
inversion becomes more difficult as the level of the input signal is raised, resulting in more
compression in the forward path.
The next stage in the inversion process can be done in one of two ways. After AGC inversion, both the positive values of the signal and the spectral extant of the signal are known.
Projections onto convex sets [11], in this case defined by the positive values of the detector output and the spectral extant of the cochlear filters, can be used to find the original
signal. This is shown in the left half of Figure 3. Alternatively, the spectral projection filter
can be combined with the next stage of processing to make the algorithm more efficient.
The increased efficiency is due to better match between the spectral projection and the
cochlear filterbank, and due to the simplified computations within each iteration. This is
shown in the right half of Figure 3. The result is an algorithm that produces nearly perfect
results with no iterations at all.
Temporal
Projection
Filter
Bank
Inversion
Spectral
Projection
~
~
~
Q)
E
AGC
-go
Inversion
Other --110.. Inversion
ChannelsL..._ _.....
~Qi
0>
CI)~
~
8
o
Figure 3. There are two ways to use convex projections to recover the information lost
by the detectors. The conventional approach is shown on the left. The right figure
shows a more efficient approach where the spectral projection has been combined with
the filterbank inversion
Finally, the multiple outputs from the cochlear filterbank are converted back into a single
waveform by correcting the phase and summing all channels. In the ideal case, each
cochlear channel contains a unique portion of the spectral energy, but with a bit of phase
delay and amplitude change. For example, if we run the signal through the same filter the
spectral content does not change much but both the phase delay and amplitude change will
be doubled. More interestingly, if we run the signal through the filter backwards, the forward and backward phase changes cancel out. After this phase correction, we can sum all
channels and get back the Original waveform, with a bit of spectral coloration. The spectral coloration or tilt can be fixed with a simple filter. A more efficient approach to correct
the spectral tilt is to scale each channel by an appropriate weight before summing, as
shown in Figure 4. The result is a perfect reconstruction, over those frequencies where the
cochlear filters are non-zero.
Figure 5 shows results from the cochleagram inversion procedure. An impulse is shown on
the left, before and after 10 iterations of the HWR inversion (using the algorithm on the
right half of Figure 3). With no iterations the result is nearly perfect, except for a bit of
noise near the center. The overall curvature of the baseline is due to the fact that informa-
830
~
Malcolm Slaney
c:UI
~'S
.... 0-
.s5
iIO
Time
Reversed
Filter for
Inversion
IIR or FIR
Filter to
Correct
Tilt
E
-0 ....
~
;:'0>
0>
CJ)~
~
c:,a
as;:,
?,e.
0>;:,
itO
Time
Reversed
Filter for
Inversion
E
-0 ....
c:.g
;:'0>
0>
CJ)~
Figure 4. Two approaches are shown here to invert the filterbank. The left diagram shows the
normal approach, the right figure shows a more efficient approach where the spectral-tilt filter
is converted to a simple multiplication.
tion near DC has been lost as it travels through the auditory system and there is no way to
recover it with the information that we have. A more interesting example is shown on the
right. Here the word "tap" 1 has been reconstructed, with and without the AGC inversion.
With the AGe inversion the result is nearly identical to the original. The auditory system
is very sensitive to onsets and quickly adapts to steady state sounds like vowels. It is interesting to compare this to the reconstruction withoutAGC inversion. Without the AGC, the
result is similar to what the ear hears, the onsets are more prominent and the vowels are
deemphasized. This is shown in the right half of Figure 5.
Impulse inversion
with no iterations
Impulse iteration
with 10 iterations
"Tap" reconstruction "Tap" Reconstruction
with AGC Inversion without AGC Inversion
Figure 5. The cochlear reconstructions of an impulse and the word "tap" are shown here. The
first and second reconstructions show an impulse reconstruction with and without iterations.
The third and fourth waveforms are the word "tap" with and without the AGe inversion.
3 CORRELOGRAM INVERSION
The correlogram is an efficient way to capture the short-time periodicities in the auditory
signal. Many mechanical measurements of the cochlea have shown that the response is
highly non-linear. As the signal level changes there are large variations in the bandwidth
and center frequency of the cochlear response. With these kinds of changes, it is difficult
to imagine a system that can make sense of the spectral profile. This is especially true for
decisions like pitch determination and sound separation.
But through all these changes in the cochlear filters, the timing information in the signal is
preserved. The spectral profile, as measured by the cochlea, might change, but the rate of
glottal pulses is preserved. Thus I believe the auditory system is based on a representation
of sound that makes short-time periodicities apparent. One such representation is the correlogram. The correlogram measures the temporal correlation within each channel, either
using FFfs which are most efficient in computer implementations, or neural delay lines
much like those found in the binaural system of the owl.
1. The syllable "tap", samples 14000 through 17000 of the "trainldr5/fcdfll
sxl06/sx106.adc" utterance on the TIMIT Speech Database, is used in all voiced
examples in this paper.
831
Pattern Playback in the '90s
The process of inverting the correlogram is simplified by noting that each autocorrelation
is related by the Fourier transform to a power spectrum. By combining many power spectrums into a picture, the result is a spectrogram. This process is shown in Figure 6. In this
way, a separate spectrogram is created for each channel. There are known techniques for
converting a spectrogram, which has amplitude information but no phase information,
back into the original waveform. The process of converting from a spectrogram back into
a waveform is described in Section 4. The correlogram inversion process consists of
inverting many spectrograms to form an estimate of a cochleagram. The cochleagram is
inverted using the techniques described in Section 2.
Frame 42 of Correlogram
Frame 43 of Correlogram
IFFTto get
of spectrogram
One line of Cochleagram
Spectrogram
Inversion .,
Time
Time
Figure 6. Correlogram inversion is possible by noting that each row of the correlogram contains the same information as a spectrogram of the same row of cochleagram output. By converting the correlogram into many spectrograms, the spectrogram inversion techniques
described in Section 4 can be used. The lower horizontal stripe in the spectrogram is due to the
narrow passband of the cochlear channel. Half-wave rectification of the cochlear filter output
causes the upper horizontal stripes.
One important improvement to the basic method is possible due to the special characteristics of the correlogram. The essence of the spectrogram inversion problem is to recover
the phase information that has been thrown away. This is an iterative procedure and would
be costly ifit had to be performed on each channel. Fortunately, there is quite a bit of overlap between cochlear channels. Thus the phase recovered from one channel can be used to
initialize the spectrogram inversion for the next channel. A difficulty with spectrogram
inversion is that the absolute phase is lost. By using the phase from one channel to initialize the next, a more consistent set of cochlear channel outputs is recovered.
4 SPECTROGRAM INVERSION
While spectrograms are not an accurate model of human perception, an implementation of
a correlogram includes the calculation of many spectrograms. Mathematically, an autocorrelation calculation is similar to a spectrogram or short-time power spectrum. One column
of a conventional spectrogram is related to an autocorrelation of a portion of the original
waveform ~y a Fourier transform (see Figure 6). Unfortunately, the final representation of
both spectrograms and autocorrelations is missing the phase information. The main task of
a spectrogram inversion algorithm is to recover a consistent estimate ofthe missing phase.
This process is not magical, it can only recover a signal that has the same magnitude spectrum as the original spectrogram. But the consistency constraint on the time evolution of
the signal power spectrum also constrains the time evolution of the spectral phase.
Malcolm Slaney
832
The basic procedure in spectrogram inversion [12] consists of iterating between the time
and the frequency domains. Starting from the frequency domain, the magnitude but not
the phase is known. As an initial guess, any phase value can be used. The individual power
spectra are inverse Fourier transformed and then summed to arrive at a single waveform.
If the original spectrogram used overlapping windows of data, the information from adjacent windows either constructively or destructively interferes to estimate a waveform. A
spectrogram of this new data is calculated, and the phase is now retained. We know the
original magnitude was correct. Thus we can estimate a better spectrogram by combining
the original magnitude information with the new phase information. It can be shown that
each iteration will reduce the error.
Figure 7 shows an outline of steps that can be used to improve the consistency of phase
estimates during the first iteration. As each portion of the waveform is added to the estimated signal, it is possible to add a linear phase so that each waveform lines up with the
proceedings segments. The algorithm described in the paragraph above assumes an initial
phase of zero. A more likely phase guess is to choose a phase that is consistent with the
existing data. The result with no iterations is a waveform that is often closer to the original
than that calculated assuming zero initial phase and ten iterations.
The total computational cost is minimized by combining these improvements with the initial phase estimates from adjacent channels of the correlogram. Thus when inverting the
first channel of the correlogram, a cross-correlation is used to pick the initial phase and a
few more iterations insure a consistent result. After the first channel, the phase of the proceeding channel is used to initialize the spectrogram inversion and only a few iterations
are necessary to fine tune the waveform.
Reconstructions from
segments 1 through N
o
Rotated Segment N+ 1
400
hi\;\; I
o
Cross Correlation
SegmentN+1
New Reconstruction
b;&/\; I
o
400
400
300
Maximize fit by
chOOSing peak
o
Figure 7. A procedure for adjusting the phase of new segments when inverting a spectrogram is
shown above. As each new segment (bottom left) is converted from a power spectrum into a
waveform, a linear phase is added to maximize the fit with the existing segments (top left.) The
amount of rotation is determined by a cross correlation (middle). Adding the new segment with
the proper rotation (top right) produces the new waveform (bottom right.)
833
Pattern Playback in the '90s
5 PUTTING IT TOGETHER
This paper has described two steps to convert a correlogram into a sound. These steps are
detailed below:
I)
For each row of the correlogram:
a)
Convert the autocorrelation data into power spectrum (Section
3).
b)
Use spectrogram inversion (Section 4) to convert the spectrograms into an estimate of cochlear channel output.
c)
Assemble the results of spectrogram inversion into an estimate
of the cochleagram.
Invert the cochleagram using the techniques described in Section 2.
2)
This process is diagrammed in Figures I and 6.
6 RESULTS
Figure 8 shows the results of the complete reconstruction process for a 200Hz impulse
train and the word "tap." In both cases, no iterations were performed for either the spectrogram or filterbank inversion. More iterations reduce the spectral error, but do not make the
graphs look better or change the perceptual quality much. It is worth noting that the "tap"
reconstruction from a correlogram looks similar to the cochleagram reconstruction without the AGC (see Figure 5.) Reducing the level of the input signal, thus reducing the
amount of compression performed by the AGC, results in a correlogram reconstruction
similar to the original waveform.
0.02 r---------.
0.05f"'"'"!~----""""
0.01
o
o
-0.01
-0.02"'--~~~~~
50 100 150 200
-0.05 " ' - - - - - - -....
o 1000 2000 3000
Figure 8. Reconstructions from the correlogram representation of an impulse train and the word
"tap" are shown above. Reducing the input signal level, thus minimizing the effect of errors
when inverting the AGe, produces results identical to the original "tap."
It is important to note that the algorithms described in this paper are designed to minimize
the error in the mean-square sense. This is a convenient mathematical definition, but it
doesn't always correlate with human perception. A trivial example of this is possible by
comparing a waveform and a copy of the waveform delayed by lOms. Using the meansquared error, the numerical error is very high yet the two waveforms are perceptually
equivalent. Despite this, the results of these algorithms based on mean-square error do
sound good.
7 CONCLUSIONS
This paper has described several techniques that allow several stages of an auditory model
to be converted back into sound. By converting each row of the correlogram into a spectrogram, the spectrogram inversion techniques of Section 4 can be used. The special characteristics of a correlogram described in Section 3 are used to make the calculation more
efficient. Finally, the cochlear filterbank can be inverted to recover the original waveform.
The results are waveforms, perceptually identical to the original waveforms.
834
Malcolm Slaney
These techniques will be especially useful as part of a sound separation system. I do not
believe that our auditory system resynthesizes partial waveforms from the auditory scene.
Yet, all research systems generate separated sounds so that we can more easily perceive
their success. More work is still needed to fine-tune these algorithm and to investigate the
ability to reconstruct sounds from partial correlograms.
Acknowledgments
I am grateful for the inspiration provided by Frank Cooper's work in the early 1950's on
pattern playback[13][14]. His work demonstrated that it was possible to convert a spectrogram, painted onto clear plastic, into sound.
This work in this paper was performed with Daniel Naar and Richard F. Lyon. We are
grateful for the help we have received from Richard Duda (San Jose State), Shihab
Shamma (U. of Maryland), Jim Boyles (The MathWorks) and Michele Covell (Interval
Research).
References
[1] Malcolm Slaney, D. Naar, R. F. Lyon, "Auditory model inversion for sound separation," Proc. of IEEE ICASSP, Volume II, pp. 77-80, 1994.
[2] M. Slaney and R. F. Lyon, "On the importance of time-A temporal representation of
sound," in Visual Representations of Speech Signals, eds. M. Cooke, S. Beet, and M.
Crawford, J. Wiley and Sons, Sussex, England, 1993.
[3] R. F. Lyon, "A computational model of binaural localization and separation," Proc. of
IEEE ICASSP, 1148-1151, 1983.
[4] M. Weintraub, "The GRASP sound separation system," Proc. of IEEE ICASSP, pp.
18A.6.1-18A.6.4, 1984.
[5] D. Hennes, "Pitch analysis," in Visual Representations of Speech Signals, eds. M.
Cooke, S. Beet, and M. Crawford, J. Wiley and Sons, Sussex, England, 1993.
[6] N. Suga, "Cortical computational maps for auditory imaging," Neural Networks, 3, 321, 1990.
[7] T. lrino, H. Kawahara, "Signal reconstruction from modified auditory wavelet transfonn," IEEE Trans. on Signal Processing, 41,3549-3554, Dec. 1993.
[8] x. Yang, K. Wang, and S. Sharnma, "Auditory representations of acoustic signals,"
IEEE Trans. on Information Theory, 38, 824-839, 1992.
[9] R. F. Lyon, "A computational model of filtering, detection, and compression in the
cochlea," Proc. of the IEEE ICASSP, 1282-1285,1982.
[10] D. Naar, "Sound resynthesis from a correlogram," San Jose State University, Department of Electrical Engineering, Technical Report #3, May 1993.
[11] R. W. Papoulis, "A new algorithm in spectral analysis and band-limited extrapolation," IEEE Trans. Circuits Sys., vol. 22, 735, 1975.
[12] D. Griffin and J. Lim, "Signal estimation from modified short-time Fourier transfonn," IEEE Trans. on Acoustics, Speech, and Signal Processing, 32, 236-242, 1984.
[13] F. S. Cooper, "Some Instrumental Aids to Research on Speech," Report on the Fourth
Annual Round Table Meeting on Linguistics and Language Teaching, Georgetown University Press, 46-53, 1953.
[14] F. S. Cooper, "Acoustics in human communications: Evolving ideas about the nature
of speech," J. Acoust. Soc. Am., 68(1),18-21, July 1980.
| 978 |@word middle:1 version:1 compression:3 inversion:45 agc:9 duda:1 cochleagram:14 instrumental:1 pulse:1 pick:1 papoulis:1 initial:5 contains:2 transfonn:2 daniel:2 interestingly:1 existing:2 recovered:3 com:1 comparing:1 yet:2 readily:1 numerical:1 designed:1 half:6 guess:2 sys:1 short:5 provides:1 mathematical:2 along:1 correlograms:1 consists:2 autocorrelation:5 paragraph:1 roughly:1 lyon:6 window:2 considering:1 becomes:1 provided:1 insure:1 alto:1 circuit:2 what:1 kind:1 spectrally:1 adc:1 acoust:1 corporation:1 temporal:3 ti:1 unwanted:1 filterbank:7 control:1 positive:3 before:2 understood:1 timing:1 engineering:1 despite:1 painted:1 firing:1 path:1 might:1 limited:2 shamma:1 suga:1 bat:1 unique:1 acknowledgment:1 lost:4 block:1 procedure:4 evolving:1 cascade:1 projection:9 convenient:1 word:5 road:1 doubled:1 get:2 onto:2 conventional:3 equivalent:2 demonstrated:1 missing:3 center:2 map:1 go:1 starting:1 convex:4 correcting:2 perceive:1 boyle:1 his:1 variation:1 controlling:1 imagine:1 stripe:2 database:1 bottom:2 wang:1 capture:1 electrical:1 ui:1 constrains:1 diagrammed:1 grateful:2 segment:7 localization:1 efficiency:1 basis:3 binaural:2 icassp:5 easily:1 train:2 separated:2 fast:2 detected:1 choosing:1 kawahara:1 apparent:1 richer:1 quite:1 plausible:1 beet:2 reconstruct:1 ability:1 transform:2 noisy:1 final:1 interferes:1 reconstruction:17 adaptation:2 relevant:1 combining:3 adapts:1 enhancement:1 produce:4 perfect:3 rotated:1 help:1 measured:1 received:1 dividing:1 implemented:1 soc:1 involves:1 judge:1 synchronized:1 waveform:24 correct:3 iio:1 filter:19 human:5 implementing:1 owl:1 mathematically:1 correction:1 normal:1 deciding:1 predict:1 early:1 estimation:1 proc:4 travel:1 palo:1 sensitive:1 successfully:1 always:1 playback:6 modified:2 improvement:2 baseline:1 am:3 sense:2 glottal:1 transformed:1 interested:1 issue:1 overall:2 raised:1 special:2 initialize:3 summed:1 identical:3 represents:1 broad:2 look:2 cancel:1 filling:1 nearly:3 minimized:1 report:2 richard:3 few:2 preserve:1 individual:1 delayed:1 phase:28 vowel:2 thrown:1 detection:2 interest:1 highly:1 investigate:1 adjust:1 grasp:1 cochleagrams:1 accurate:1 closer:1 partial:2 necessary:1 increased:1 column:1 modeling:1 cost:1 delay:3 iir:1 my:1 combined:2 peak:1 sensitivity:1 destructively:1 together:1 quickly:1 extant:2 ear:1 choose:1 fir:1 slaney:8 converted:7 includes:1 onset:2 piece:1 multiplicative:1 performed:6 view:1 tion:1 extrapolation:1 portion:5 wave:3 recover:6 option:1 slope:1 voiced:1 timit:1 minimize:1 square:2 characteristic:2 ofthe:1 covell:1 plastic:1 accurately:1 multiplying:1 worth:2 apple:1 history:2 detector:3 ed:2 definition:1 energy:2 echolocation:1 frequency:4 pp:2 weintraub:1 gain:3 auditory:20 adjusting:1 lim:1 cj:2 amplitude:3 back:7 response:2 improved:1 done:1 stage:6 correlation:4 horizontal:2 overlapping:1 autocorrelations:1 quality:2 indicated:1 michele:1 perhaps:1 believe:2 impulse:7 effect:1 true:1 evolution:2 inspiration:1 adjacent:2 round:1 during:1 essence:1 speaker:3 steady:1 sussex:2 prominent:1 outline:1 complete:1 instantaneous:1 rotation:2 tilt:4 volume:1 s5:1 measurement:1 automatic:1 consistency:2 teaching:1 nonlinearity:1 language:1 had:1 add:1 curvature:1 own:1 success:1 meeting:1 inverted:4 fortunately:1 spectrogram:39 employed:1 converting:4 maximize:2 signal:26 ii:1 july:1 multiple:1 sound:26 technical:1 match:1 england:2 plausibility:1 cross:3 determination:1 calculation:4 divided:1 equally:1 controlled:1 qi:1 pitch:3 basic:2 fffs:1 iteration:17 represent:1 cochlea:4 invert:2 dec:1 preserved:2 background:2 fine:2 interval:3 addressed:3 diagram:2 hz:1 near:2 yang:2 backwards:2 intermediate:1 split:1 noting:4 ideal:1 fit:2 bandwidth:1 reduce:2 idea:1 motivated:1 speech:8 cause:1 useful:2 iterating:1 detailed:2 clear:1 ifit:1 tune:2 amount:2 band:3 ten:1 generate:1 estimated:1 track:1 vol:1 putting:1 salient:1 backward:1 imaging:1 graph:1 sum:1 convert:4 run:2 inverse:1 jose:2 powerful:1 fourth:2 extends:1 arrive:1 separation:9 decision:1 summarizes:1 griffin:1 bit:4 hi:1 syllable:1 display:1 assemble:1 annual:1 constraint:1 scene:1 aspect:1 fourier:4 performing:1 department:1 combination:1 describes:3 son:2 rectification:1 mathworks:1 needed:1 know:1 available:1 apply:1 away:1 appropriate:2 spectral:20 voice:1 deemphasized:1 original:15 magical:1 assumes:1 running:1 top:2 linguistics:1 especially:3 passband:1 added:2 costly:1 reversed:2 separate:2 maryland:1 cochlear:23 trivial:1 assuming:1 retained:1 minimizing:1 difficult:3 unfortunately:1 frank:1 negative:1 constructively:1 implementation:2 proper:1 conversion:2 upper:1 communication:1 jim:1 dc:1 frame:2 inverting:5 mechanical:1 tap:10 meansquared:1 acoustic:4 narrow:1 trans:4 below:1 pattern:7 perception:4 lpc:1 including:1 power:7 critical:1 overlap:1 difficulty:1 improve:1 movie:1 picture:1 created:1 utterance:1 hears:1 crawford:2 understanding:1 georgetown:1 multiplication:1 loss:1 neurophysiologically:1 interesting:3 filtering:1 age:8 consistent:4 bank:3 interfering:1 cooke:2 row:4 periodicity:3 copy:1 loms:1 allow:3 absolute:1 calculated:2 cortical:1 doesn:1 forward:2 san:2 simplified:2 ec:1 correlate:1 reconstructed:1 summing:2 consuming:1 alternatively:1 spectrum:8 iterative:1 table:1 channel:24 nature:1 ca:1 domain:3 psychoacoustic:2 main:1 arrow:1 noise:2 profile:2 cooper:3 wiley:2 aid:1 position:1 perceptual:3 coloration:2 third:1 ito:1 wavelet:1 rectifier:1 explored:1 adding:1 importance:1 ci:1 magnitude:4 perceptually:5 suited:1 mill:1 simply:1 likely:1 neurophysiological:3 visual:2 correlogram:37 tracking:1 informa:1 content:1 change:9 determined:1 except:1 reducing:3 reversing:1 engineer:1 called:1 total:1 pas:1 evaluate:2 malcolm:8 resynthesis:1 |
7,219 | 979 | Learning in large linear perceptrons and
why the thermodynamic limit is relevant
to the real world
Peter Sollich
Department of Physics, University of Edinburgh
Edinburgh EH9 3JZ, U.K.
P.Sollich~ed.ac.uk
Abstract
We present a new method for obtaining the response function 9
and its average G from which most of the properties of learning
and generalization in linear perceptrons can be derived. We first
rederive the known results for the 'thermodynamic limit' of infinite
perceptron size N and show explicitly that 9 is self-averaging in
this limit. We then discuss extensions of our method to more general learning scenarios with anisotropic teacher space priors, input
distributions, and weight decay terms. Finally, we use our method
to calculate the finite N corrections of order 1/ N to G and discuss
the corresponding finite size effects on generalization and learning
dynamics. An important spin-off is the observation that results
obtained in the thermodynamic limit are often directly relevant to
systems of fairly modest, 'real-world' sizes.
1
INTRODUCTION
One of the main areas of research within the Neural Networks community is the issue
of learning and generalization. Starting from a set of training examples (normally
assumed to be input-output pairs) generated by some unknown 'teacher' rule V, one
wants to find, using a suitable learning or training algorithm, a student N (read
'Neural Network') which generalizes from the training set, i.e., predicts the outputs
corresponding to inputs not contained in the training set as accurately as possible.
208
Peter Sollich
If the inputs are N-dimensional vectors x E nN and the outputs are scalars yEn,
then one of the simplest functional forms that can be assumed for the student .N is
the linear perceptron, which is parametrized in terms of a weight vector W N E nN
and implements the linear input-output mapping
YN(X) = 7Nw~x.
(1)
A commonly used learning algorithm for the linear perceptron is gradient descent
on the training error, i.e., the error that the student .N makes on the training set .
Using the standard squared output deviation error measure, the training error for a
given set ofp training examples {(xlJ , ylJ),J-l = 1 . . . p} is E t = L:IJ ~(ylJ-YN(XIJ))2 =
~ L:IJ (ylJ - w~xlJ / VN)2. To prevent the student from fitting noise in the training
data, a quadratic weight decay term ~AW~ is normally added to the training error,
with the value of the weight decay parameter A determining how strongly large
weight vectors are penalized. Gradient descent is thus performed on the function
E
E t + ~AW~, and the corresponding learning dynamics is, in a continuous time
approximation , dw N / dt = -"V wE. As discussed in detail by Krogh and Hertz
(1992), this results in an exponential approach of W N to its asymptotic value , with
decay constants given by the eigenvalues of the matrix M N , defined by (1 denotes
the N x N identity matrix)
=
MN
= Al + A,
A
= ~ L:IJ xlJ(xlJ)T.
To examine what generalization performance is achieved by the above learning
algorithm, one has to make an assumption about the functional form of the teacher .
The simplest such assumption is that the problem is learnable, i. e., that the teacher ,
like the student, is a linear perceptron. A teacher V is then specified by a weight
vector Wv and maps a given input x to the output Yv(x) = w~ x/VN. We assume
that the test inputs for which the student is asked to predict the corresponding
outputs are drawn from an isotropic Gaussian distribution , P(x) <X exp( _~x2). The
generalization error, i. e., the average error that a student .N makes on a random
input when compared to teacher V, is given by
fg
= ~((YN(X) -
yv(x))2)p(x)
= 2~(WN -
w v )2.
(2)
Inserting the learning dynamics W N = W N(t), the generalization acquires a time
dependence, which in its exact form depends on the specific training set , teacher, and
initial value of the student weight vector, WN(t = 0). We shall confine our attention
to the average of this time-dependent generalization error over all possible training
sets and teachers; to avoid clutter , we write this average simply as fg(t). We assume
that the inputs x lJ in the training set are chosen independently and randomly from
the same distribution as the test inputs, and that the corresponding training outputs
are the teacher outputs corrupted by additive noise, ylJ = yv(xlJ) + 171J, where the 171J
have zero mean and variance rr2. If we further assume an isotropic Gaussian prior on
the teacher weight vectors, P(w v ) <X exp( -~w~), then the average generalization
error for t -> 00 is (Krogh and Hertz, 1992)
fg(t ->
1 [ rr 2 C + A(rr2 - A) OC]
00) = 2
OA '
(3)
where G is the average of the so-called response function over the training inputs:
C = (Q) P( {x"}),
Q --
N1 t r
1
MN?
(4)
Learning in Large Linear Perceptrons
209
The time dependence of the average generalization error for finite but large t is an
exponential approach to the asymptotic value (3) with decay constant oX + amin,
where am in is the lowest eigenvalue occurring in the average eigenvalue spectrum of
the input correlation matrix A (Krogh and Hertz, 1992). This average eigenvalue
spectrum, which we denote by p(a), can be calculated from the average response
function according to (Krogh, 1992)
p(a) =
!11' f-O+
lim ImGI.>.=_a-if,
(5)
where we have assumed p( a) to be normalized, fda p( a) = 1.
Eqs. (3,5) show that the key quantity determining learning and generalization in
the linear percept ron is the average response function G defined in (4). This
function has previously been calculated in the 'thermodynamic limit', N -- 00
at a = piN = const., using a diagrammatic expansion (Hertz et al., 1989) and the
replica method (Opper, 1989, Kinzel and Opper, 1991). In Section 2, we present
what we believe to be a much simpler method for calculating G , based only on
simple matrix identities. We also show explicitly that 9 is self-averaging in the
thermodynamic limit, which means that the fluctuations of 9 around its average G
become vanishingly small as N -> 00 . This implies, for example , that the generalization error is also self-averaging. In Section 3 we extend the method to more
general cases such as anisotropic teacher space priors and input distributions, and
general quadratic penalty terms. Finite size effects are considered in Section 4,
where we calculate the O(IIN) corrections to G, ?g(t -- (0) and p(a) . We discuss
the resulting effects on generalization and learning dynamics and derive explicit conditions on the perceptron size N for results obtained in the thermodynamic limit
to be valid. We conclude in Section 5 with a brief summary and discussion of our
results.
2
THE BASIC METHOD
Our method for calculating the average response function G is based on a recursion
relation relating the values of the (unaveraged) response function 9 for p and p + 1
training examples. Assume that we are given a set of p training examples with
corresponding matrix M N . By adding a new training example with input x, we
obtain the matrix Mt = MN + ~ xx T . It is straightforward to show that the
inverse of Mt can be expressed as
1 M-1
TM-1
(MN+) - 1 -_ MN-1 - IV N XX 1N
1 + N1 X TM-N x
.
(One way of proving this identi~ is to multiply both sides by Mt .and exploit ~he
fact that MtM~',l = 1 + kxx M;~} . ) TaklOg the trace, we obtalO the follow 109
recursion relation for g:
1 ..!.x?-2x
(6)
9 (p + 1) = 9 (p) - N 1 N 1 T ~ -1 .
+NX
N X
Now denote Zi = ~xTMNix (i = 1,2). With x drawn randomly from the assumed
input distribution P(x) ex: exp( - ~x2), the Zi can readily be shown to be random
210
Peter Sollich
variables with means and (co-)variances
(A
A)
2
M-i-j
~Zi~Zj = N1 tr
N
?
Combining this with the fact that tr M:vk ~ N>.-k = O(N), we have that the
(Zi)
=
1
M-i
N tr
N'
fluctuations LlZi of the
into (6), we obtain
Zi
around their average values are 0(1/.JN); inserting this
9(p + 1)
l I t M- 2
9(p) _ N r N
N 1 + ~trM:vl
+ 0(N-3/2)
9(p) + ~ a9(p)
1
+ 0(N-3/2).
(7)
N ()>. 1 + 9(p)
Starting from 9(0)
1/>., we can apply this recursion p
aN times to obtain
3
9(p) up to terms which add up to at most 0(pN- /2) = 0(1/.JN). This shows
that 9 is self-averaging in the thermodynamic limit: whatever the training set, the
value of 9 will always be the same up to fluctuations of 0(1/.JN). In fact, we shall
show in Section 4 that the fluctuations of 9 are only 0(1/ N). This means that the
0(N- 3 / 2 ) fluctuations from each iteration of (7) are only weakly correlated, so that
they add up like independent random variables to give a total fluctuation for 9(p)
of O?p/ N 3 )1/2) 0(1/ N).
=
=
=
We have seen that, in the thermodynamic limit, 9 is identical to its average G
because its fluctuations are vanishingly small. To calculate the value of G in the
thermodynamic limit as a function of a and >., we insert the relation 9(p+1)-9(p) =
-ka9(a)/aa + 0(1/ N 2 ) into eq. (7) (with 9 replaced by G) and neglect all finite N
corrections. This yields the partial differential equation
aG aG 1
aa - a>. 1 + G = 0,
(8)
which can readily be solved using the method of characteristic curves (see, e.g.,
John, 1978). Using the initial condition Glo:o = 1/>. gives a/(l + G) = l/G - >.,
which leads to the well-known result (see, e.g., Hertz et al., 1989)
;>. (1 - a - >. + J(l - a - >.)2 + 4>') .
(9)
In the complex >. plane, G has a pole at >. = 0 and a branch cut arising from
G=
the root; according to eq. (5), these singularities determine the average eigenvalue
spectrum pea) of A, with the result (Krogh, 1992)
pea)
= (1- a)8(1 -
a)6(a)
1 J(a+
+ -2
1ra
- a)(a - a_),
(10)
where 8(x) is the Heaviside step function, 8(x) = 1 for x > 0 and 0 otherwise.
The root in eq. (10) only contributes when its argument is non-negative, i.e., for a
(1 ? fo)2.
between the 'spectral limits' a_ and a+, which have the values a?
=
3
EXTENSIONS TO MORE GENERAL LEARNING
SCENARIOS
We now discuss some extensions of our method to more general learning scenarios. First, consider the case of an anisotropic teacher space prior, P(w v ) ex:
Learning in Large Linear Perceptrons
211
exp(-!w~:E~lWV)' with symmetric positive definite :E v . This leaves the defini-
tion of the response function unchanged; eq. (3), however, has to be replaced by
1/2{q 2G + A[q2 - A(~tr :Ev)]oGloA}.
fg(t -- 00)
=
As a second extension, assume that the inputs are drawn from an anisotropic distribution, P(x) oc exp(-~xT:E-1x). It can then be shown that the asymptotic value
of the average generalization error is still given by eq. (3) if the response function is
redefined to be 9
~ tr :EM~l. This modified response function can be calculated
as follows: First we rewrite 9 as ~tr (A:E- 1 + A)-I, where A = ~ L:JJ(xJJ)Tx JJ is
the correlation matrix of the transformed input examples xl' = :E- 1/ 2x JJ . Since the
x JJ are distributed according to P(x JJ ) oc exp( _~(xJJ)2), the problem is thus reduced
to finding the response function 9 = ~ tr (L + A)-1 for isotropically distributed
inputs and L
A:E- 1. The recursion relations between 9(p + 1) and 9(p) derived
in the previous section remain valid, and result, in the thermodynamic limit, in a
?i!f~rential ~~uat.ion for the aver are response function ~ analo.gous. t? eq. (8). The
1mt1al cond1tlon 1S now Gla=o = N tr L -1, and one obtams an 1mphc1t equatiOn for
=
=
G,
1 (
G = N tr L
a
+ 1+G 1
)-1 ,
(11)
where in the case of an anisotropic input distribution considered here, L = A:E- 1.
If :E has a particularly simple form, then the dependence of G on a and A can be
obtained analytically, but in general eq. (11) has to solved numerically.
Finally, one can also investigate the effect of a general quadratic weight decay term,
!W;:AWN' in the energy function E. The expression for the average generalization
error becomes more cumbersome than eq. (3) in this case, but the result can still be
expressed in terms of the average response function G
(9) = (~tr (A + A)-l),
which can be obtained as the solution of eq. (11) for L = A .
=
4
FINITE N CORRECTIONS
So far, we have focussed attention on the thermodynamic limit of perceptrons of
infinite size N. The results are clearly only approximately valid for real, finite
systems, and it is therefore interesting to investigate corrections for finite N. This
we do in the present section by calculating the O(IIN) corrections to G and pea).
For details of the calculations and results of computer simulations which support
our theoretical analysis, we refer the reader to (Sollich, 1994).
=
First note that, for A
0, the exact result for the average response function is
GI.~=o = (a - 1 - IIN)-l for a > 1 + liN (see, e.g., Eaton, 1983), which clearly
admits a series expansion in powers of liN. We assume that a similar expansion
also exists for nonzero A, and write
(12)
Go is the value of G in the thermodynamic limit as given by eq. (9). For finite N, the
fluctuations ~9 = 9-G of9 around its average value G become relevant; for A = 0,
the variance of these fluctuations is known to have a power series expansion in II N,
and again we assume a similar expansion for finite A, ((~9)2) = ~2 IN + 0(11 N2),
212
Peter Sollich
where the first term is 0(1/ N) and not 0(1) because, as discussed in Section 2, the
fluctuations ofg for large N are no greater than 0(1/v'N). To calculate G I and A 2 ,
one starts again from the recursion relation (6), now expanding everything up to
second order in the fluctuation quantities AZi and Ag. Averaging over the training
inputs and collecting orders of l/N yields after some straightforward algebra the
known eq. (8) for Go and two linear partial differential equations for G I and A 2 ,
the latter obtained by squaring both sides of eq . (6) . Solving these , one obtains
G _ G6(1 - AGO)
I (1 + AG6)2
(13)
=
In the limit A -- 0, G 1
l/(a - 1)2 consistent with the exact result for G quoted
above; likewise, the result A 2 == 0 agrees with the exact series expansion of the
variance of the fluctuations of 9 for A = 0, which begins with an 0(1/ N 2 ) term
(see, e.g., Barber et ai., 1994).
(a)
(b)
0.5
1.5
- - A=O.OOl
0.4
-------. A
=0.1
1.0
0.3
0.5
0.2
0.0
0.1
-0.5
0.0
0.0
0.5
1.0
a
1.5
2.0
0.5
0.0
1.0 a
1.5
2.0
Figure 1: Average generalization error: Result for N -- 00 , (g,O, and coefficient of
O(l/N) correction, (g , l. (a) Noise free teacher, (72 O. (b) Noisy teacher, (72 0 .5.
Curves are labeled by the value of the weight decay parameter A.
=
=
From the l/N expansion (12) of G we obtain , using eq. (3), a corresponding expansion of the asymptotic value of the average generalization error, which we write
as (g(t -- 00)
(g ,O + (g , d N + 0(1/ N2). It follows that the thermodynamic limit
result for the average generalization error , (g ,O, is a good approximation to the true
result for finite N as long as N ~ Nc = /fg ,I/(g ,ol. In Figure 1, we plot (g,O and
(g,l for several values of A and (j 2 . It can be seen that the relative size of the first
order correction l(g,I/(g,ol and hence the critical system size Nc for validity of the
thermodynamic limit result is largest when A is small. Exploiting this fact, Nc can
be bounded by 1/(1 - a) for a < 1 and (3a + l)/[a(a - l)J for a > l. It follows,
for example , that the critical system size Nc is smaller than 5 as long as a < 0.8
or a > l.72 , for all A and (j2. This bound on Nc can be tightened for non-zero
A; for A > 2, for example , one has Nc < (2A - l)/(A + 1)2 < 1/3. We have thus
shown explicitly that thermodynamic limit calculations of learning and generalization behaviour can be relevant for fairly small , ' real-world' systems of size N of the
order of a few tens or hundreds. This is in contrast to the widespread suspicion
=
Learning in Large Linear Perceptrons
213
among non-physicists that the methods of statistical physics give valid results only
for huge system sizes of the order of N :::::: 10 23 .
(b)
(a)
po
PI
I
(1 - 0-)0(1 - 0-)
4
I
4
a
a
a __- - - - _
Figure 2: Schematic plot of the average eigenvalue spectrum p( a) of the input
correlation matrix A. (a) Result for N -+ 00, po(a). (b) O(l/N) correction, Pl(a).
Arrows indicate 6-peaks and are labeled by the corresponding heights.
We now consider the 0(1/ N) correction to the average eigenvalue spectrum of the
input correlation matrix A. Setting p(a)
po(a) + pda)/ N + 0(1/ N 2 ), po(a) is
the N -+ 00 result given by eq. (10), and from eq . (13) one derives
=
Figure 2 shows sketches of Po (a) and PI (a). Note that fda PI (a) = 0 as expected
since the normalization of p( a) is independent of N. Furthermore, there is no
O(l/N) correction to the 6-peak in po(a) at a = 0, since this peak arises from the
N - p zero eigenvalues of A for c.r = p/ N < 1 and therefore has a height of 1- c.r for
any finite N. The 6-peaks in PI (a) at the spectral limits a+ and a_ are an artefact
of the truncated l/N expansion: p(a) is determined by the singularities of G as
a function of A, and the location of these singularities is only obtained correctly
by resumming the full l/N expansion. The 6-peaks in pl(a) can be interpreted
as 'precursors' of a broadening of the eigenvalue spectrum of A to values which,
when the whole 1/ N series is resummed, will lie outside the N -+ 00 spectral
range [a_ , a+]. The negative term in PI (a) represents the corresponding 'flattening'
of the eigenvalue spectrum between a_ and a+ . We can thus conclude that the
average eigenvalue spectrum of A for finite N will be broader than for N -+ 00,
which means in particular that the learning dynamics will be slowed down since the
smallest eigenvalue amin of A will be smaller than a_. From our result for PI (a) we
can also deduce when the N -+ 00 result po(a) is valid for finite N; the condition
turns out to be N ~ a/[(a+ - a)(a - a_)]. Consistent with our discussion of the
broadening of the eigenvalue spectrum of A, N has to be larger for a near the
spectral limits a_, a+ if po(a) is to be a good approximation to the finite N average
eigenvalue spectrum of A.
214
5
Peter Sollich
SUMMARY AND DISCUSSION
We have presented a new method, based on simple matrix identities, for calculating
the response function 9 and its average G which determine most of the properties
of learning and generalization in linear perceptrons. In the thermodynamic limit,
N --+ 00, we have recovered the known result for G and have shown explicitly that 9
is self-averaging. Extensions of our method to more general learning scenarios have
also been discussed. Finally, we have obtained the 0(1/ N) corrections to G and the
corresponding corrections to the average generalization error, and shown explicitly
that the results obtained in the thermodynamic limit can be valid for fairly small,
'real-world' system sizes N . We have also calculated the 0(1/N) correction to the
average eigenvalue spectrum of the input correlation matrix A and interpreted it
in terms of a broadening of the spectrum for finite N, which will cause a slowing
down of the learning dynamics.
We remark that the O( 1/N) corrections that we have obtained can also be used
in different contexts, for example for calculations of test error fluctuations and
optimal test set size (Barber et al., 1994). Another application is in an analysis of
the evidence procedure in Bayesian inference for finite N, where optimal values of
'hyperparameters' like the weight decay parameter A are determined on the basis of
the training data (G Marion , in preparation). We hope, therefore, that our results
will pave the way for a systematic investigation of finite size effects in learning and
generalization.
References
D Barber, D Saad, and P Sollich (1994). Finite size effects and optimal test set size
in linear perceptrons. Submitted to J. Phys. A.
M LEaton (1983). Multivariate Statistics - A Vector Space Approach. Wiley, New
York.
J A Hertz, A Krogh, and G I Thorbergsson (1989) . Phase transitions in simple
learning. J. Phys. A, 22:2133-2150.
F John (1978). Partial Differential Equations. Springer, New York, 3rd ed.
W Kinzel and M Opper (1991). Dynamics of learning. In E Domany, J L van Hemmen, and K Schulten, editors, Models of Neural Networks, pages 149-171. Springer,
Berlin.
A Krogh (1992). Learning with noise in a linear perceptron. J. Phys. A, 25:11191133.
A Krogh and J A Hertz (1992). Generalization in a linear percept ron in the presence
of noise. J. Phys. A, 25:1135-1147.
M Opper (1989). Learning in neural networks: Solvable dynamics. Europhysics
Letters, 8:389-392.
P Sollich (1994). Finite-size effects in learning and generalization in linear perceptrons. J. Phys. A, 27:7771-7784.
| 979 |@word simulation:1 tr:10 initial:2 series:4 recovered:1 readily:2 john:2 additive:1 plot:2 leaf:1 slowing:1 plane:1 isotropic:2 location:1 ron:2 simpler:1 height:2 become:2 differential:3 fitting:1 expected:1 ra:1 examine:1 ol:2 precursor:1 becomes:1 begin:1 xx:2 bounded:1 lowest:1 what:2 interpreted:2 q2:1 ag:3 finding:1 collecting:1 uk:1 whatever:1 normally:2 yn:3 positive:1 limit:22 physicist:1 fluctuation:13 approximately:1 co:1 range:1 implement:1 definite:1 procedure:1 area:1 context:1 map:1 straightforward:2 attention:2 starting:2 independently:1 go:2 rule:1 dw:1 proving:1 exact:4 particularly:1 cut:1 predicts:1 labeled:2 solved:2 calculate:4 pda:1 asked:1 defini:1 dynamic:8 weakly:1 rewrite:1 solving:1 algebra:1 basis:1 po:8 tx:1 outside:1 larger:1 otherwise:1 statistic:1 gi:1 noisy:1 a9:1 eigenvalue:15 rr:1 vanishingly:2 inserting:2 relevant:4 combining:1 j2:1 amin:2 exploiting:1 derive:1 ac:1 ij:3 ex:2 eq:16 krogh:8 implies:1 indicate:1 artefact:1 pea:3 everything:1 behaviour:1 azi:1 generalization:23 investigation:1 of9:1 singularity:3 extension:5 insert:1 correction:15 pl:2 confine:1 around:3 considered:2 exp:6 mapping:1 nw:1 predict:1 eaton:1 smallest:1 agrees:1 largest:1 hope:1 clearly:2 gaussian:2 always:1 modified:1 mtm:1 avoid:1 pn:1 broader:1 derived:2 ylj:4 vk:1 contrast:1 am:1 inference:1 dependent:1 squaring:1 nn:2 vl:1 lj:1 relation:5 transformed:1 issue:1 among:1 unaveraged:1 fairly:3 identical:1 represents:1 few:1 randomly:2 replaced:2 phase:1 n1:3 huge:1 investigate:2 multiply:1 gla:1 partial:3 modest:1 iv:1 kxx:1 theoretical:1 pole:1 deviation:1 marion:1 hundred:1 teacher:14 aw:2 corrupted:1 peak:5 systematic:1 physic:2 off:1 squared:1 again:2 student:8 coefficient:1 explicitly:5 depends:1 performed:1 root:2 tion:1 yv:3 start:1 yen:1 spin:1 variance:4 characteristic:1 percept:2 likewise:1 yield:2 bayesian:1 accurately:1 ago:1 submitted:1 fo:1 cumbersome:1 phys:5 ed:2 energy:1 lim:1 dt:1 follow:1 response:14 ox:1 strongly:1 trm:1 furthermore:1 correlation:5 sketch:1 widespread:1 believe:1 effect:7 validity:1 normalized:1 ofg:1 true:1 analytically:1 hence:1 read:1 symmetric:1 nonzero:1 self:5 acquires:1 oc:3 iin:3 functional:2 kinzel:2 mt:3 anisotropic:5 discussed:3 extend:1 he:1 relating:1 numerically:1 refer:1 ai:1 rd:1 glo:1 add:2 deduce:1 multivariate:1 scenario:4 wv:1 seen:2 greater:1 determine:2 ii:1 branch:1 thermodynamic:17 full:1 calculation:3 long:2 lin:2 ofp:1 europhysics:1 schematic:1 basic:1 iteration:1 normalization:1 achieved:1 ion:1 want:1 saad:1 near:1 presence:1 wn:2 zi:5 tm:2 domany:1 expression:1 penalty:1 ool:1 peter:5 xjj:2 cause:1 jj:5 remark:1 york:2 clutter:1 ten:1 simplest:2 reduced:1 xij:1 zj:1 arising:1 correctly:1 write:3 shall:2 key:1 drawn:3 prevent:1 replica:1 inverse:1 letter:1 reader:1 vn:2 eh9:1 bound:1 quadratic:3 x2:2 fda:2 awn:1 argument:1 department:1 according:3 hertz:7 remain:1 sollich:9 xlj:5 em:1 smaller:2 slowed:1 g6:1 equation:4 previously:1 discus:4 pin:1 turn:1 generalizes:1 gous:1 apply:1 spectral:4 jn:3 denotes:1 const:1 calculating:4 neglect:1 exploit:1 unchanged:1 added:1 quantity:2 dependence:3 pave:1 gradient:2 berlin:1 oa:1 parametrized:1 nx:1 barber:3 nc:6 trace:1 negative:2 redefined:1 unknown:1 observation:1 finite:20 descent:2 truncated:1 community:1 pair:1 specified:1 identi:1 ev:1 power:2 suitable:1 critical:2 solvable:1 recursion:5 mn:5 brief:1 suspicion:1 prior:4 determining:2 asymptotic:4 relative:1 interesting:1 consistent:2 tightened:1 editor:1 pi:6 penalized:1 summary:2 free:1 side:2 perceptron:6 focussed:1 fg:5 edinburgh:2 distributed:2 curve:2 calculated:4 opper:4 world:4 valid:6 transition:1 van:1 commonly:1 far:1 resumming:1 obtains:1 conclude:2 assumed:4 a_:8 quoted:1 spectrum:12 continuous:1 why:1 jz:1 expanding:1 obtaining:1 contributes:1 expansion:10 broadening:3 complex:1 flattening:1 main:1 arrow:1 whole:1 noise:5 hyperparameters:1 n2:2 hemmen:1 wiley:1 rr2:2 schulten:1 explicit:1 exponential:2 xl:1 lie:1 rederive:1 uat:1 aver:1 down:2 diagrammatic:1 specific:1 xt:1 learnable:1 decay:8 admits:1 evidence:1 derives:1 exists:1 adding:1 occurring:1 simply:1 expressed:2 contained:1 scalar:1 isotropically:1 springer:2 aa:2 thorbergsson:1 identity:3 infinite:2 determined:2 averaging:6 called:1 total:1 perceptrons:9 support:1 latter:1 arises:1 preparation:1 heaviside:1 correlated:1 |
7,220 | 98 | 553
SPREADING ACTIVATION OVER
DISTRIBUTED MICROFEATURES
James Hendler *
Depart.ment, of Computer Science
University of Maryland
College Park, MD 20742
ABSTRACT
One att?empt at explaining human inferencing is that of spreading activat,ion, particularly in the st.ructured connectionist paradigm. This has resulted in t.he building of systems with semantically nameable nodes which perform inferencing by examining
t.he pat,t.erns of activation spread. In this paper we demonst.rate
t.hat simple structured network infert'ncing can be p(>rformed by
passing art.iva.t.ion over the weights learned by a distributed algarit,hm. Thus , an account, is provided which explains a wellbehaved rela t ionship bet.ween structured and distri butt'd conn('ct.ionist. a.pproachrs.
INTRODUCTION
A primar~? difference brtween t,he nPllral net.works of 20 years ago and t.he
("urrent genera Lion of connect,ionist models is t.he addit.ion of mechanisms whic h
permit t.he s),st,em to create all internal represent,ation. These subsymbolic,
semantica.lly unnameable, feat.urrs which a.re induced by connectionist. learning
algorithms havr been discussed as bt,ing of import. bot,h in structured and distribut.ed ronnl"ctionist nrtworks (cf. Feldman and Balla.rd , 1982; Rumelhart and
McClelland, 198(j). The fact that network learning algorit.hms can creatr these
rm?cro!eal'ure,s is not, however. enough in itself t.o aC('Qunt for how rognition
works. Most. of what, we call int.elligent thought. dt'rives from being able t,o rea:son about. t.he relatioll:::> I)t'tween object.s, to hypothesize about event.s a.nd things,
etc. If we are to do cognit.ive modeling we must. complet.e the story by rxplainiJlg
how networks can rea,-;oll in tht> wa.y that, humans (or other int,elligrnt beings) do.
aile attempt at (-'xplaining such rea.<;oning is that of spreading activat.ion ill the
structured cOllllect,ionist. and marker- passing (cf. Charniak, ]983; Hendler, 1987)
?
The aut.hor is also affiliatf' ci wit.h thp Instit.ut.e for Acivanced C'omputpr Studies a.nd t.he Systems Research Center a.t the Universit.y of Ma.ryland. Funding for this work was provided in part by
Offi ce of Naval Resf'arch Grallt N00014-88--K - 0,560 .
554
Hendler
approaches. In t,hese syst,em::i semantically nanwable nodes permit an energy
spread; and reasoning about tilt' world is accounted for by looking at, either
stahh' configurations of tJw activation (the st.ruet.ured connectionist. approach) or
at. t,he paths fOllnd by examining int.ersect,ions among t.he nodes (the markerpassing te('llnique) . In this paper we will demonst.rate t.hat. simple Rt,ruct.lJrednetwork- like infE'rencing ('an be performed by pa.ssing act.ivat.ion over t,he wt'ightR
lea.rned by a distribut.ed algorithm. Thus, an account. is provided which explains
a well-behaved relation~hip bet.ween st.ruct,ured and dist.ributed connt'ct.ionist
a pproa.ches .
THE SPREADING ACTIVATION MODEL
In this paper we will demonstrate that local connect.ionist.- Iike net.works can be
built by spreading activat.ion ovel' t.he microfeaturcs learned b.\' a dist.ribut.ed network. To show t.his, we st,art wit.h a simple example which c1f'Jl1onst.rat.es the
activation spreading mechanism used. ThE' part.icular net.work We will use ill this
example is a 6- 3-8 three-layer net.work t rained by t.he hack-propagation learning
algorithm. The training set used is shown in table 1. The \w?ights bet.ween the
out.put node::; and hidden units which are learned by the Iletwork (after h'arning
t.o tlw !)O% level for a typical rlln) are shown in figure 1 .
TABLE 1. Training Set. for Examph' 1.
Inpll t
Pattern
Output.
Pattern
000000
0000 J 1
001100
OOll}1
110000
10000000
01000000
00100000
00010000
00001000
00000100
00000010
00000001
lIOO]]
] I I 100
111111
Spreading Activation over Distributed Microfeatures
Weights
n1
n2
n3
n4
n5
n6
n7
nB
h1
h2
h3
-4.98
-6.99
-6.11
-6.37
4.36
4.38
0.89
3.88
4.40
-4.99
3.49
-4.68
3.73
-5.97
1.07
-6.95
-2.82
-2.23
0.30
2.53
-5.09
-3.67
3.32
1.88
Figure 1. Weights L~arnecl by Back Propagation
To underst.and how t,he act,ivat.ioll spreads, let. us examine what. occurs when
activation is started at node nl wit.h a weight of 1. This activation strength is
divided by the outbranching of the nodf' and then mult.iplied by t.he weight of
each link to t.he hidden units. Thus a.ctiva.tion flows from nl t.o hi with a
strength of 1/ 3 1" Weighl(tl1 ,hl}. A similar computat.ion is made t.o each of t.he
other hidden units. This act.ivat.ion now spreads to each of the ot.her Ollt.pUt.
nodes in turn. Thus, n2 would gain act,ivat.ion of
Acti','ation(hl) I Wet"ght(n2,hl)/ 8 -1Activat-ion(h2) 1" lYe'ight(n2)d)/ 8 +
Ad1'vation(h3) 1" Weight(n2,h3)/ 8
or .80 from rli .
Table 2 shows a graph of t.he act.ivat,ioll spread between the output units. The
table, which is symmetric, can t.hus be read as showing t.he out,put at each of the
other units when an activation st.rength of 1 is placed at t,he named node . Looking at. the table we see t,hat. t.il(' hight:'st, activat.ion OCClJrs among nodes which
share t.he most feat,ures of tJlf' input (i .e. ~anlf' value and posit.ion) while t.he
lowest is se('n among tho::;t:' pat.tt'fIls shariug t.he fewest feat.ures.
555
556
Hendler
However, as well as having this property, t.able 2 can be seen as providing a
matrix which specifies t.he weights bet.ween t.he out.put nodes if viewed as a st.ruct,ured net.work. That. is. ,,1 is conned?ed to rtf! by a strength of + .80, to nB by a
st.rength of + 1.0:3, etc . Thus. by I\~ing this t.echniquE' distribut.ed r{'presentations
can be turned into connectivity weight.s for st.ructured lIet.works. When non-?
ort.hogonal wpights are lIsed. t he same act.ivat ion--spreading a.lgorithm produces a
struct.ured network whi('h can be used for more complex inferenciug than can the
dist.ributed net work alone .
We demonst.rat.e t.his b)' a simple . and aga.in contrin?d. example. This example is
mot.ivat.ed by Gary Cot.trell's st.ructured JI10dd for word sense disambiguation
(CoUrt'll, Hl85). Cottrell, using weights clerived by hand, demonstrat.ed that. a
struct.ured connect.ionis!. net.work could di:';t inguish both word- sense a.nd case- slot
assignm('nts for amhiguous lexical items. Pre~enkd wit.h the sent{'nce "John
threw the fight." f?he syst.em would ndivate a node for one meaning of "throw/'
presented wit.h "John t.hrew t.he ball" it would come up wit.h another. The nodes
of Cottrell's network included worels (John. Threw , etc .). word RenseR (Johnl,
Propp\. etc .) :lnd cClI'P- sloLs (TAGT (lIgt'llt of tl1(-' throw) . PAGT (ag('nt of t,he
Propel). etc.).
TABLE 2. Adivat.jon Spn>ad in 6--:3?8 Network.
,,1
nl
n2
n.'J
n4
'115
116
n7
uB
?
.80
1.03
.17
.38
-1.57
- .38
- 2.3
II::!
.80
oj
?
1.03
1.02
-.14
.97
- .63
-2 .03
- .03
-l.f!7
1.02
2.60
-1.57
.31
-.79
?
114
?//5
.17
2.6
.97
?
-2.42
-.38
-.09
.52
.38
-1.57
- .0:3
- 2.42
?
.64
-.;38
- .77
uB
-1.57
.:31
- 2.03
-.:38
.04
?
- .f)
2.1 ?1
Tl7
- .:38
-.79
- .0:3
-.Of!
-.;38
-.6
?
.09
u8
- 2.3
.14
--1.97
.52
-.77
2 .14
.09
?
To duplicate Gary's network via training, we presented a 3-layer backprop net...
work wit.h a training set in which distribut.ed pat.t.erns, H'ry loosely corresponding
to a "dictionary" of word (-,Ileodings l were associa.ted wit,h a vector representing
each of the individual noell's which would he represented in Cottrt-ll's system, but.
wit.h no struct.ure . Thus, each elemPIlt. in t.he training set. is
}- Whieh in any realisti r sy stem would sornp. day be rpplaeed by aduaJ signal pro.;pssing outp"t.S Of Othpf rpprespnt.alinn s of ai'liial worn pfonuneiat.ioll forms.
Spreading Activation over Distributed Microfeatures
a 16 bit vretor (represent.ing a four word srnt.en('e. ea('h word as a. 4 bit. pattrrn),
associated wit.h allot.he)? 16 bit. ve('tor repr<'spnt.ing t.he nodes
Bobl Johnl propE'l t.hrfW IJght I ball J jlJgt pob) t agt. to bj bob John t.hrfw U1E'
fight ball
For t.his example, the system was t.rnineci on th(? E'1I('odings of t.he four srnt.encrs
John t.hrew the ball
John threw the fight.
Bob t.1Hfw t be hall
Bob t.hrfw Lllf 6ght
wit.h the output set. high foJ' tho:-;e objects
a.ppropriatf?\y a.<.;so('iaj,rd . (IS shown in Tabh? :3.
TABLE
;~.
the second vedor whi('h \v('/,e
Tr<lining S(?t fa)' Example Z.
Output.
Patt.el'n
Input.
Pattern
011 0
01 J 0
100/
J 00 J
/II
0001 0101
OOOJ 0101
OOOJ 010J
000 1 0101
OOlO
J010
00/0
10 10
JOO 11000 1J J 01 I 10
lOlOOlJI00101101
0/0/100011011110
01 JOOJ 1100011101
Upon complet.ion of t.he kaming , t.he aet.ivat.ioJl sprrading algorithm was used to
derive a table of COlll1f'ctivity weights betw('ell the out.put. unit.s as shown in table
4.
These weight.s W(')'e then t.ransferred into a local COllnect ionist simulat.or and a
very simple act.ivation spn'adiJlg modrl was llsed to examine t.he r esult?s . \Vhpn
Wt' run t.he simulator. u~ing tht' aeti\'atioIl spn?ading oY('r leaJ'lwd wpights.
exactly t.he f('sult.s prodllced by Cott.reJl's n{'twork lIr{' seen . Thus :
Act.ivation from tht, nodt's corresponding to john. tbroll', tllf-. alld fif/hl
cam,e a posit.ive activation at. the node for ('Throw" and a negative a('t.ivat.ion at t.he node for "Propr!."
while
A('tivat.ion from joh" throlt' the ball sprt'ad positiniy to " Prope'" ,lIlel
not. t.o 'I t IIrow . ,.
Furt.her, ot?her effrct.s which
cHP
also predict.ed by C'ou.rdl's model lire seeJl:
Act.iYat.ioJl at. TAGT and TOB.! spreads posit.i\'t' activation to TIr.,.o/l'
and not t.o Propel.
and
Activation at PAG?' and POB) causes a spread t.o Propel but. not to
Thro'W.
557
558
Hendler
TABLE 4. Connectivity Weights for Example 2.
*** -?0.12
- 0.12 ***
0.01 -OJ)}
-0 .01 0.01
-0.0] 0.01
0,01 -0,01
O.OJ -0.01
0 .01 ?-0.01
-0.01 0.01
-0.01 0.01
0 .12 -0.12
- 0.12 0.12
-0.0:3 0.03
??0.03 0 .0:3
??0.0] 0.01
0.00 O.OJ
0.01 -0.01
- 0.01 0.01
*u -0.04
?-0.04 ***
-0.04 0.04
0,04 -0.04
0.04 -0.04
0.05 -0.05
-0,05 0.05
-0.04 0.04
0 .01 -0.01
?-0,01 0 .01
-0.02 0 .02
??0.02 0 ,02
-?0.04 0.04
0.04 -0.04
-0.01 0.01
0.01 - 0.01
-0.04 0 .04
0.04 -0 .04
u* -0.04
-0,04 ***
-0,05 O.O!)
-0,05 0.05
0.05 -0.05
0.04 -0.04
- 0.01 0.01
0.01 -?0.01
0.02 -?0.02
0.02 -0 .02
0.04 - 0.04
?-0.04 0.04
0.01
-0.01
0 .04
- 0.04
-0.0,5
0.05
***
O.OS
-0.05
-0,05
0.01
-0.00
-?0.02
?-0.02
-?0.04
0.0,5
0.01 - 0.01 -0.01
-0.01 0.01 0.01
0.05 -0.05 - 0.04
-0.05 0.05 0.04
-0,05 0.05 0,04
0,05 -0.0.5 -? 0.04
0.05 -0 .05 -0.05
u* -?0.05 -O.O!)
- 0.05 *** 0.05
-0.05 0.05 u*
0 .01 -0.01 -0 .01
-0.01 0 .01 O.OJ
-0.02 0.02 0.02
-0.02 0 .0;3 0.02
- 0.04 O.O<J D.O?'
0,05 O.O!) O.OG
0.12 - 0 .12
-?0.12 0 .12
0.01 - 0.01
?_?0.01 0.01
-0,01 0 ,01
0.01 - 0.01
0.01 ??0.00
0.01 -O.OJ
-0.01 O.()]
-0.01 0.01
*** ?-O.l:.?
??0.12 ***
?-0.0:3 0.0:3
0.0:3 0.0:)
? 0.01 O.OJ
O.OJ ?0.00
???0.0:3
0.0:3
- 0 .02
0.02
0.02
- 0 .02
-0.02
-0'()2
0 .02
0.02
?-0.0:3
O.O:~
***
0.20
0.02
0 .02
-0.0:3 -0.01 0 .00
O.O~ 0.01 ?-0.01
- 0.02 - 0.04 0.04
0.02 0.04 - 0 .04
0.02 0.04 - 0.04
-0.02 - 0.04 0.04
-0.02 - 0 .04 0.05
-0,02 ??0.04 0.05
0.0:3 0 .04 - 0.05
0.02 0.04 -0.05
-0.0:3 ?-0.01 0.0]
0.0:3 0 .01 ??0.00
0 .20 0.02 -0.02
* ** 0.02 - 0.02
0.02 *** . ~.O .?
0.02 ?0.04 ***
We believe that results like this one ma.y argue t.hat. :-;tl'uetllrt'd IId?works are
int.pgrally linked to dist.ributed networks ill t,hat. distribllt.<?d Ilf'twork le:ullillg
t.echuiqut's may provide a fundamental ba~i:-; for ('xplaining the ('ogllit.in' df'VelOPment. of structured networks . III addition, Wt ' see that :,-imple illf{'J'f'llt ial rt'a~on?
ing ('an be produ('ed using purdy cOIlJledioni~t. mockl:,.
CONCLUDING REMARKS
Wt' have at.t.empt.ed t.o ;;;hO\.... that a model u;;;ing an activation spn'ading va.rinnt
ean be used t.o take learned connect.ionist models aJld lwrform SOllle limit ed forms
of inferencing upon t.helll. Furt.her, WP have argllt"d t.hat t.his tf'chniqlll' Illay pro"idr a comput.ational modd in which strllctured network;;; C;}l) Iw leal'llPd and
t.hat st,ruct.ured net.work::; provide the infeI't'ncing (?apnbilities missing in purely
dist.rihut.ed models. However. befort' w(' can t.ruly furtht'r thi~ claim. ~ignit1('ant
work remain;;; to be done . We must ext.end alld ('xplort' such models. particularly
t'xamining whether t.hese t?ypes of t.echniqups ('an be extt'llded t.o handle tlw complt'xitr t?hat can be found in real?- world prol)JpIll~ and sl'rioll~ ('ogllitin models .
In part ieula.r we are beginlling an examinat.ion of two cru('ial isslIP::;: First.,
will t.he technique described above work for realist.ic problem?;? In particular, can
t.hl' infen?ncing be designed t.o impact on t.1l(' ]'Pwgnition by t 11(> di.stribut.ed net.work? If so, one ('ould Sf'f', for l'xample, a speech recognit.ion program coupled to
a system like Cottrell's natural languagt' sy:.;\.eJll, providing a handle for a t.ext.
underst.anding system. Similarly such a t.e('hnique might. allow the int.('gration of
top - down and bott.OIll'-IIP pro('('ssing for vision a.nd ot.ht'r such signal Pl'ocf'ssing
tasks.
Spreading Activation over Distributed Microfeatures
Secondly. we wish to see jf more complex spreading activation models could be
hooked to this t.ype of model. Could networks such as t.hose proposed by Sha.st.ri
(1985), Diedt'rich (HI8!?). and Polla.ck and 'Valt.z (1982) which provide complex
inferencing bllt requirt' more ,structure than simply wt'ight.s bet.ween unit.s, be
ab.stract.t'cI out. of the learned weight.s? Two partieular areas current.ly being pur~llf'd by t.JH' author. for t'xampk foclls on act.i\'e inhibit.ioll modeb for dt't.ermining whetlwr port.jon~ of the nt'twork C(ln lw ~lIpPJ'f'ssed t.o provide mort' complex
inferencing and t.he learning of structures given temporally ordered information.
References
Charniak, E. Passing markt'rs: A t.heory of contt'xtual influence
comprehension Cognitive S6ence, 7(3), 198:~. 171-190.
lJl
language
Cottrell, G.W. A Connection-ist Approach to Word - Sense DiMlmb1'gllation, Doet.oral Disst'rta.t.ion, Comput.t'r Scienc(' Dt'pt.. l fllivl'rsit.y of Hochesler, 1985.
Diederich, J. Parallelverarbeil/l'flg in 1IetzlL'erk-ba.~ie1'fen Systemen PhD Disst'rtation, Dt'pt. of Linguistics, llni\'ersity of Rielt~f('ld. J985.
Feldman, J.A. and Ballard , D.H. (J 982). Connect.ionist models a.nd t.heir propt:'rt.ies. Cognit1've Science. 6. 205-?254-
JlItegrating Maf'l.?er -pali8illg aod Problem Solving: A .~p,.e(/dirtg
artivatiou approach to 1'ml)ro'l'cd choice ill ploflJlillg Lawrence Erlbaulll As~ociate~,
N .J., Novt'mbt'J', Hlg7.
Hendler, J.A.
Pollack, J.B. and 'Yalt.7., D.L Nat?u!'al Language Processing lIsiug spr?'ading
act.ivat.ion and lateral inhibit.ion Proceedings of the FO'llrth intema.i1'o'l1al Conference of Ihe Cognitive S'o'eure S'ociety, 1982, .50-58.
Rurnelhart . D.E. and McClelland. J.L. ((,ds.) Parallel D';strib'llfed Computing
Cambridge . ,Ma.: MJT Prel-is.
Shastri, L. El'ideul1'al RWMuing ?in SefJ/(llItic /\.retll'{)rk.~: A formal theor:1J ((tid //8
parallel implementatioTl Doct.oral DiS8t'l't at.jon, Computt'r Scit'nce Depart ment,
l Jnivefsit.y of Roc-hester, Sept .. J 985.
559
| 98 |@word joh:1 underst:2 llsed:1 nd:5 r:1 tr:1 fif:1 ld:1 configuration:1 cru:1 att:1 charniak:2 current:1 nt:3 activation:16 import:1 must:2 john:7 cottrell:4 wellbehaved:1 hypothesize:1 designed:1 heir:1 alone:1 item:1 scienc:1 ial:2 tjw:1 cognit:1 node:14 acti:1 idr:1 cot:1 dist:5 examine:2 ry:1 simulator:1 provided:3 distri:1 lowest:1 what:2 erk:1 prope:2 iva:1 ag:1 act:11 exactly:1 ro:1 universit:1 rm:1 pur:1 unit:7 ly:1 tid:1 local:2 limit:1 ext:2 propp:1 ure:2 path:1 might:1 genus:1 tir:1 lira:1 area:1 thi:1 thought:1 mult:1 chp:1 word:7 pre:1 hight:1 put:5 nb:2 influence:1 ilf:1 demonstrat:1 center:1 lexical:1 missing:1 worn:1 wit:11 his:4 handle:2 tht:3 pt:2 anding:1 pa:1 rumelhart:1 hose:1 particularly:2 algorit:1 cro:1 ssing:3 inhibit:2 imple:1 cam:1 maf:1 hese:2 solving:1 oral:2 rtf:1 purely:1 upon:2 represented:1 hooked:1 fewest:1 recognit:1 whi:2 ive:2 simulat:1 tlw:2 itself:1 net:10 ment:2 turned:1 heory:1 produce:1 object:2 derive:1 inferencing:5 ac:1 h3:3 throw:3 come:1 posit:3 human:2 sult:1 explains:2 backprop:1 secondly:1 comprehension:1 theor:1 tho:2 pl:1 hall:1 ic:1 aga:1 lawrence:1 bj:1 predict:1 claim:1 tor:1 dictionary:1 threw:3 wet:1 spreading:11 iw:1 create:1 iojl:2 tf:1 ck:1 bet:5 og:1 naval:1 sense:3 el:2 vedor:1 purdy:1 bt:1 fight:3 hidden:3 relation:1 her:4 i1:1 ocf:1 among:3 ill:4 distribut:4 art:2 ell:1 having:1 ted:1 ribut:1 weighl:1 park:1 jon:3 tl1:2 connectionist:3 duplicate:1 resulted:1 ve:2 individual:1 n1:1 attempt:1 ab:1 propel:3 nl:3 llf:1 furt:2 hester:1 loosely:1 re:1 ollt:1 ivat:10 pollack:1 hip:1 eal:1 modeling:1 xample:1 examining:2 vation:1 mot:1 connect:5 st:13 oolo:1 ured:6 fundamental:1 ches:1 connectivity:2 iip:1 mjt:1 allot:1 activat:5 illf:1 cognitive:2 syst:2 account:2 int:5 ad:2 performed:1 h1:1 tion:1 linked:1 iike:1 parallel:2 thp:1 alld:2 il:1 sy:2 ant:1 iid:1 bob:3 ago:1 lwd:1 llt:2 fo:1 ed:14 diederich:1 energy:1 james:1 hack:1 associated:1 di:2 gain:1 ut:1 ou:1 ea:1 back:1 dt:4 day:1 iaj:1 done:1 arch:1 d:1 hand:1 o:1 marker:1 propagation:2 behaved:1 hor:1 believe:1 building:1 lgorithm:1 read:1 symmetric:1 wp:1 ll:2 ivation:2 echnique:1 rat:2 tt:1 demonstrate:1 pro:3 reasoning:1 meaning:1 demonst:3 funding:1 tilt:1 discussed:1 he:42 cambridge:1 feldman:2 ai:1 rd:2 similarly:1 repr:1 language:2 hus:1 joo:1 etc:5 ort:1 n00014:1 fen:1 seen:2 aut:1 rained:1 paradigm:1 ween:5 ight:2 signal:2 ii:2 stem:1 ing:7 divided:1 infe:1 y:1 va:1 impact:1 n5:1 vision:1 df:1 lly:1 represent:2 ion:22 lea:1 rea:3 addition:1 ures:2 twork:3 ionist:8 ot:3 induced:1 sent:1 thing:1 flow:1 n7:2 call:1 microfeatures:4 oill:1 iii:1 enough:1 ioll:4 court:1 ould:1 rli:1 whether:1 rurnelhart:1 modd:1 ie1:1 mbt:1 passing:3 cause:1 speech:1 remark:1 hi8:1 complet:2 se:1 mcclelland:2 instit:1 specifies:1 sl:1 computat:1 bot:1 patt:1 icular:1 pob:2 conn:1 four:2 ist:1 ce:1 ht:1 graph:1 year:1 nce:2 run:1 nameable:1 named:1 aile:1 ctivity:1 disambiguation:1 bit:3 layer:2 ct:2 hi:1 strength:3 n3:1 ri:1 whic:1 concluding:1 bllt:1 ern:2 structured:5 ball:5 associa:1 remain:1 em:3 son:1 n4:2 hl:5 ypes:1 aet:1 ln:1 turn:1 mechanism:2 ihe:1 end:1 permit:2 mort:1 ype:1 struct:3 ho:1 hat:8 top:1 cf:2 linguistics:1 spn:4 ght:2 rdl:1 outp:1 depart:2 occurs:1 fa:1 sha:1 rt:3 md:1 link:1 maryland:1 lateral:1 addit:1 argue:1 ruct:4 providing:2 shastri:1 negative:1 ba:2 sprt:1 perform:1 pat:3 looking:2 connection:1 learned:5 able:2 foj:1 lion:1 pattern:3 program:1 built:1 oj:8 oooj:2 event:1 ation:2 natural:1 representing:1 pproa:1 temporally:1 started:1 hendler:6 hm:2 rela:1 n6:1 coupled:1 sept:1 lining:1 oy:1 h2:2 port:1 story:1 share:1 cd:1 accounted:1 placed:1 formal:1 allow:1 jh:1 explaining:1 pag:1 distributed:5 leal:1 lye:1 world:2 rich:1 author:1 made:1 ights:1 feat:3 lir:1 ml:1 butt:1 arning:1 ljl:1 table:10 ballard:1 ean:1 rta:1 spr:1 complex:4 tween:1 spread:7 n2:6 tob:1 en:1 tl:1 roc:1 ational:1 ovel:1 wish:1 sf:1 comput:2 lw:1 subsymbolic:1 down:1 rk:1 hogonal:1 flg:1 showing:1 er:1 oll:1 ad1:1 ci:2 agt:1 phd:1 aod:1 te:1 nat:1 simply:1 ordered:1 gary:2 ma:3 slot:1 viewed:1 presentation:1 u8:1 jf:1 markt:1 included:1 typical:1 semantically:2 wt:5 e:1 college:1 internal:1 betw:1 ub:2 tllf:1 |
7,221 | 980 | Interior Point Implementations of
Alternating Minimization Training
Michael Lemmon
Dept . of Electrical Engineering
University of Notre Dame
Notre Dame, IN 46556
lemmon@maddog.ee.nd.edu
Peter T. Szymanski
Dept. of Electrical Engineering
University of Notre Dame
Notre Dame, IN 46556
pszymans@maddog.ee.nd.edu
Abstract
This paper presents an alternating minimization (AM) algorithm
used in the training of radial basis function and linear regressor
networks. The algorithm is a modification of a small-step interior
point method used in solving primal linear programs. The algorithm has a convergence rate of O( fo,L) iterations where n is a
measure of the network size and L is a measure of the resulting
solution's accuracy. Two results are presented that specify how
aggressively the two steps of the AM may be pursued to ensure
convergence of each step of the alternating minimization.
1
Introduction
In recent years, considerable research has investigated the use of alternating minimization (AM) techniques in the supervised training of radial basis function
networks. AM techniques were first introduced in soft-competitive learning algorithms[l]. This training procedure was later shown to be closely related to
Expectation-Maximization algorithms used by the statistical estimation community[2]. Alternating minimizations search for optimal network weights by breaking
the search into two distinct minimization problems. A given network performance
functional is extremalized first with respect to one set of network weights and then
with respect to the remaining weights. These learning procedures have found applications in the training of local expert systems [3], and in Boltzmann machine
training [4]. More recently, convergence rates have been derived by viewing the AM
Michael Lemmon. Peter T. Szymanski
570
method as a variable metric algorithm [5].
This paper examines AM as a perturbed linear programming (LP) problem. Recent
advances in the application of barrier function methods to LP problems have resulted in the development of "path following" or "interior point" (IP) algorithms [6].
These algorithms are characterized by a fast convergence rate that scales in a sublinear manner with problem size. This paper shows how a small-step IP algorithm
for solving a primal LP problem can be modified into an AM training procedure ..
The principal results of the paper are bounds on how aggressively the legs of the
alternating minimization can be pursued so that the AM algorithm maintains the
sublinear convergence rate characteristic of its LP counterpart and so that both legs
converge to an optimal solution.
2
Problem Statement
Consider a function approximation problem where a stochastic approximator learns
a mapping f : IRN -+ IR. The approximator computes a predicted output, Y E IR,
given the input z E IRN. The prediction is computed using a finite set of M
regressors. The mth regressor is characterized by the pair (Om,w m) E IRN X IRN
(m 1, ... , M). The output of the mth regressor, Ym E IR, in response to an input,
z E IRN is given by the linear function
=
Ym
=O~z.
(1)
=
The mth regressor (m
1, ... , M) has an associated radial basis function (RBF)
with parameter vector Wm E IRN. The mth RBF weights the contribution of the
mth output in computing Y and is defined as a normal probability density function
Q(mlz) = km exp(-u- 2 1I wm - zW)
(2)
where k m is a normalizing constant. The set of all weights or gating probabilities is
denoted by Q. The parameterization of the mth regressor is em (O~ ,w~)T E IR2N
(m = 1, ... , M) and the parameterization of the set of M linear regressors is
=
e --
(eT
I " " , eT)T
M
.
(3)
The preceding stochastic approximator can be viewed as a neural network. The
network consists of M + 1 neurons. M of the neurons are agent neurons while the
other neuron is referred to as a gating neuron. The mth a~ent neuron is parameterized by Om, the first element of the pair em = (O?'n, w?'n) (m = 1, ... , M). The
agent neurons receive as input, the vector z E IRN. The output of the mth agent
neuron in response to an input z is Ym O?'nz (m 1, ... , M). The gating neuron is
parameterized by the conditional gating probabilities, Q. The gating probabilities
are defined by the set of vectors, Ii; = {WI, ... ,WM }. The gating neuron receives the
agent neurons' outputs and the vector z as its input. The gating neuron computes
the network's output, Y, as a hard (4) or soft (5) choice
=
Y = Ym; m = arg
A
Y=
=
max
m=I .... ,M
Q(mlz)
L:~-I Q(mlz)Ym
M
L:m=1 Q(mlz)
(4)
(5)
Interior Point Implementations of Alternating Minimization Training
571
The network will be said to be "optimal" with respect to a training set T = {( Zi, Yi) :
Yi = f(zd, i = 1, ... , R} if a mean square error criterion is minimized. Define the
square output error of the mth agent neuron to be em(zi)
(Yi - O~Zi)2 and the
square weighting or classifier error of the mth RBF to be Cm(Zi) Ilw m - Zi 112. Let
the combined square approximation error of the m,th neuron be dm (Zi) = Kee m (Zi) +
KcCm(Zi) and let the average square approximation error of the network be
=
M
d(Q, e, T) =
L
=
R
LP(zi)Q(mlzddm(zi)'
(6)
m=l i=l
Minimizing (6) corresponds to minimizing both the output error in the M linear
regressors and the classifier error in assigning inputs to the M regressors.: Since T
is a discrete set and the Q are gating probabilities, the minimization of d(Q, e, T)
is constrained so that Q is a valid conditional probability mass function over the
training set, T.
Network training can therefore be viewed as a constrained optimization problem.
In particular, this optimization problem can be expressed in a form very similar
to conventional LP problems. The following notational conventions are adopted to
highlight this connection. Let x E lR,MR be the gating neuron's weight vector where
(7)
=
Let em
(O~,w~)T E lR,2N denote the parameter vectors associated with the
regressor and define the cost vector conditioned on e = (er, ... , et)T as
c(e)
= (p(zt}d1(zt), ... ,p(zR)d1(ZR),p(Z2)d 2(Z2), ... ,p(zi)dm(Zi), ...)T
mth
(8)
With this notation, the network training problem can be stated as follows,
mmlmlze
cT(e)x
with respect to
x, e
subject to Ax = b, x 2:: 0
where b = (1, ... ,1? E lR,R, A
Xi 2:: 0 for i = 1, ... , MR.
= [IRxR? .. IRxR]
(9)
E IR RXMR , and x 2:: 0 implies
One approach for solving this problem is to break up the optimization into two
steps. The first step involves minimizing the above cost functional with respect to
x assuming a fixed e. This is the Q-update of the algorithm. The second leg of the
algorithm minimizes the functional with respect to e assuming fixed x. This leg
is called the e-update. Because the proposed optimization alternates between two
different subsets of weights, this training procedure is often referred to as alternating
minimization. Note that the Q-update is an LP problem while the e-update is a
quadratic programming problem. Consequently, the AM training procedure can be
viewed as a perturbed LP problem.
3
Proposed Training Algorithm
The preceding section noted that network training can be viewed as a perturbed
LP problem. This observation is significant for there exist very efficient LP solvers
572
Michael Lemmon, Peter T. Szymanski
based on barrier function methods used in non-linear optimization. Recent advances
in path following or interior point (IP) methods have developed LP solvers which
exhibit convergence rates which scale in a sublinear way with problem size [6]. This
section introduces a modification of a small-step primal IP algorithm that can be
used for neural network training. The proposed modification is later shown to
preserve the computational efficiency enjoyed by its LP counterpart.
To see how such a modification might arise, we first need to examine path following
LP solvers. Consider the following LP problem.
cTx
mmlmlze
with respect to
x E IRn
subject to Ax = b, x 2: 0
(10)
This problem can be solved by solving a sequence of augmented optimization problems arising from the primal parameterization of the LP problem.
mmlmlze
with respect to
subject to
O'O:)cTx(k) -
2:i logx~k)
x(k) E IRn
Ax(k) = b, x(k) 2: 0
(11)
where a(k) 2: 0 (k = 1,? .. ,1<) is a finite length, monotone increasing sequence of
real numbers. x*(a(k? denotes the solution for the kth optimization problem in
the sequence and is referred to as a central point. The locus of all points, x* (O'( k?
where a(k) 2: 0 is called the central path. The augmented problem takes the original
LP cost function and adds a logarithmic barrier which keeps the central point away
from the boundaries of the feasible set. As a increases, the effect of the barrier
is decreased, thereby allowing the kth central point to approach the LP problem's
solution in a controlled manner.
Path following (IP) methods solve the LP problem by approximately solving
the sequence of augmented problems shown in (11). The parameter sequence,
0'(0),0'(1), ... , a(K), is chosen to be a monotone increasing sequence so that the
central points, x*(a(k?, of each augmented optimization approach the LP solution
in a monotone manner. It has been shown that for specific choices of the a sequence,
that the sequence of approximate central points will converge to an f-neighborhood
of the LP solution after a finite number of iterations. For primal IP algorithms, the
required condition is that successive approximations of the central points lie within
the region of quadratic convergence for a scaling steepest descent (SSD) algorithm
[6]. In particular, it has been shown that if the kth approximating solution is sufficiently close to the kth central point and if O'(k+ 1 ) O'(k)(l + v/..{ii) where v ~ 0.1
controls the distance between successive central points, then the "closeness" to the
(k + lyt central point is guaranteed and the resulting algorithm will converge in
O( foL) iterations where L = n + p+ 1 specifies the size of the LP problem and p is
the total number of bits used to represent the data in A , b, and c. If the algorithm
takes small steps, then it is guaranteed to converge efficiently.
=
The preceding discussion reveals that a key component to a path following method's
computational efficiency lies in controlling the iteration so that successive central
points lie within the SSD algorithm's region of quadratic convergence. If we are to
successfully extend such methods to (9), then this "closeness" of successive solutions
must be preserved by the 6-update of the algorithm. Due to the quadratic nature
Interior Point Implementations of Alternating Minimization Training
573
of the e-update, this minimization can be done exactly using a single NewtonRaphson iteration. Let e* denote e-update's minimizer. If we update e to e*,
it is quite possible that the cost vector, c(e), will be rotated in such a way that
the current solution, x(k), no longer lies in the region of quadratic convergence.
Therefore, if we are to preserve the IP method's computational efficiency it will be
necessary to be less "aggressive" in the e-update. In particular, this paper proposes
the following convex combination as the e-update
e~+l) = (1 - "Y(k?e~) + "Y(k)e~+l),*
(12)
where e~) is the mth parameter vector at time k and 0 < "Y(k)
of the update. This will ensure convergence of the Q-update.
< 1 controls the size
Convergence of the AM algorithm also requires convergence of the e-update. For
the e-update to converge, "Y(k) in (12) must go to unity as k increases. Convergence
of "Y(k) to unity requires that the sequence Ile~+l),* -e~)11 be monotone decreasing.
As the e-update minimizer, e(k+l),*, depends upon the current weights, Q(k)(mlz),
large changes to Q can prevent the sequence from being monotone decreasing. Thus,
it is necessary to also be less "aggressive" in the Q-update. An appropriate bound
on v is the proposed solution to guarantee convergence of the e-update.
Algorithm 1 (Proposed Training Algorithm)
Initialize
k = O.
Choose X~k). e~). and O'(k) for i= 1,???,(MR). and for m= 1,???,M.
repeat
O'(k+l) = O'(k)(1 + v/Vn). where v ~ 0.1.
Q-update:
= X(k)
for i = 0, ... , P - 1
Xo
Xi+l
X(k+l)
= ScalingSteepestD.escent(x~k+l), O'(k+l), e(k?)
= xp
e-update:
For m
= 1, ... , M
e~+l)
= (1 - r(k?)e~)
+ r(k)e~+l),*
k=k+l
until(.6. < c)
4
Theoretical Results
This section provides bounds on the parameter, ,(k), controlling the AM algorithm's
e-update so that successive x(k) vector solutions lie within the SSD algorithm's
region of quadratic convergence and on v controlling the Q-update so that successive
central points are not too far apart, thus allowing convergence of the e-update.
Theorem 1 Let e~) and e~ ),* be the current and minimizing parameter vecc(e(k? and c(k),*
c(e(k),*). Let 6(x, a, e)
tors at time k. Let c(k)
=
=
=
574
Michael Lemmon. Peter T. Szymanski
IIPAXX (O'c(8) - x-I) II be the step size of the SSD update where PA = 1AT (AAT)-l A and X = diag(xl, ... , x n ) . Assume that 6(x CHl ), O'CHl), 8 Ck ) =
61 < 0.5 and let e~+1) = (1 - ')'Ck)e~) + ')'Ck)e~+l), ?. If ')'Ck) is chosen as
Ck) <
62 - 61
- n(I + 1I/y'n)llcCHl),. - C<k)11
where 61 < 62 = 0.5 , then 6(x(Hl), O'(Hl), 8 CH1 ) ~ 62 = 0.5.
(13)
')'
Proof: The proof must show that the choice of ')'(k) maintains the nearness ofx(k+l)
to the central path after the 8-update. Let h(x,0' , 8) = PAxX(O'c(8)-x-l),
hI = h(x CH1 ), O'CHl), 8 Ck ) and h2 = h(x(Hl) , O'(Hl), 8(Hl). Using the triangle
inequality produces
IIh211 ~ IIh2 - hIli + Ilhlll?
Ilhlll = 61 by assumption, so
Ilh 211::; 1100(Hl)PAXX(c(Hl) - c(k)11
+ 61.
Using the convexity of the cost vectors produces (c(Hl) - c(k) ::; (1 - ')'(k)c(k)
,),(k )c Ck +1) ,. - c(k) resulting in
IIh211 ~ II00CHl),),Ck)PAXX(cCHl),. - c(k)11
Using the fact that IIPAXXII ~ ~
IIh211
~
~
+
+ 61 .
= n/O'(k) (~ is the duality gap),
,),Ck)O'(Hl)IIPAXXllllc(Hl),. - c(k)11 + 61.
')'(k)n(I + 1I/y'n)llc(Hl),. - cCk)11 + 61.
Plugging in the value of ')'(k) from (13) and simplifying produces the desired result
IIh211 ~ 62 ~ 0.5, guaranteeing that x(Hl) remains close to x(Hl),. after the 8update.
0
Theorem 1 shows that the non-linear optimization can be embedded within the steps
of the path following algorithm without it taking solutions too far from successive
central points. The following two results, found in [7], provide a bound on II to
guarantee convergence of the 8-update. The bound on II forces successive central
points to be close and ensures convergence of the 8-update.
Proposition 1 Let B = LzpzQ(mlz)zzT , E = Lzpz 6Q(mlz)zzT , w =
Lz pzQ(mlz)y(z)z, and 6w = Lz p z 6Q(mlz)y(z)z, where Q(mlz) = Q(k)(mlz)
and 6Q(mlz) = Q(Hl)(mlz) - Q(Ck)mlz). Assume that ly(z)1 ~ Y, llzll ::; (, and
that B is of full rank for all valid Q 'So Finally, let J1.max = sUPQ IIB- II. Then,
118~+1),. - 8~)'?11 ~ 2(112
where I(
= J1.maxY( (1 + Cr!'ru )
and r
+ 211)K
(14)
= IIB-lIIIIEII < 1.
Theorem 2 Assume that Ilzll ~ (, ly(z)1 ~ Y, 118m ll ~ 8 max , and that B =
Lz pzQ(mlz)zzT is of full rank for all valid Q's and that liB-III IIEII = r < 1. If
II
< min {O.I,
-1 + ..jI7"I-+-r-;-/(=2"'7>(2:-J1.-m-ax""") ,
-1
+ JI + ')'min f e/(2K)}
(15)
Interior Point Implementations of Alternating Minimization Training
575
=
where 1\ = ILmax Y?l + (2 ILmax/(l- r?, 1min
(6 2 - 6d/(n(1 + 0.1/ y'n)llc(l),? c(O)11) and f0 is the largest lIer!+1),? - er!)11 such that 1(.1:) = 1, then the e-update
will converge with Iler!+1),? - er!)II-+ 0 and 1(.1:) -+ 1 as k increases.
The preceding results guarantee the convergence of the component minimizations
separately. Convergence of the total algorithm relies on the simultaneous convergence of both steps. This is currently being addressed using contraction mapping
concepts and stability results from nonlinear stability analysis [8].
The convergence rate of the algorithm is established using the LP problem's duality
gap. The duality gap is the difference between the current solutions for the primal
and dual formulations of the LP problem. Path following algorithms allow the
duality gap to be expressed as follows
A(
L..).
(k? _ n
a
-
+ 0.5y'n
a(k)
(16)
.
and thus provide a convenient stopping criterion for the algorithm . Note that
a(O) /{3 k where {3 ~ (1+11/ y'n). This implies that ~(k)
13k ~(O) ~ {3k2L. If k
is chosen so that {3k2L :::; 2- L , then ~(k) :::; 2- L which implies that k ~ 2L/ log(l/{3).
Inserting our choice of {3 one finds that k ~ (2y'nL/II)+2L. The preceding argument
establishes that the .proposed convergence rate of O( y'nL) iterations. In other
words, the procedure's training time scales in a sublinear manner with network size.
=
a(k)
=
5
Simulation Example
Simulations were run on a time series prediction task to test the proposed algorithm.
The training set is T = {(Zi,Yi) : Yi = y(iT),Zi = (Yi-l,Yi-2, ... ,Yi-Nf E lR,N}
for i = 0,1, ... ,100, N = 4, and T = 0.04 where the time series is defined as
y(t) = sin(1I"t) - sin(211"t) + sin(311"t) - sin(1I"t/2)
(17)
The results describe the convergence of the algorithm. These experiments consisted of 100 randomly chosen samples with N 4 and a number of agent neurons
ranging from M
4 to 20. This corresponds to an LP problem dimension of
404 to 2020. The stopping criteria for the tests was to run until the solution
n
was within f = 10- 3 of a local minimum. The number of iterations and floating
point operations (FLOPS) for the AM algorithm to converge are shown in Figures l(a) and l(b) with AM results denoted by "0" and the theoretical rates by a
solid line . The algorithm exhibits approximately O( y'nL) iterations to converge as
predicted. The computational cost, however is O(n 2 L) FLOPS which is better than
the predicted O( n 3 .5 L). The difference is due to the use of sparse matrix techniques
which reduce the number of computations. The resulting AM algorithm then has
the complexity of a matrix multiplication instead of a matrix inversion. The use of
the algorithm resulted in networks having mean square errors on the order of 10- 3 .
=
6
=
=
Discussion
This paper has presented an AM algorithm which can be proven to converge in
O( foL) iterations. The work has established a means by which IP methods can be
576
Michael Lemmon, Peter T. Szymanski
104r---------~--________~
G)
~
1015 ,....-_ _ _ _ _ _ _ _- - - - - - - - - - - ,
o
~
>
~G)
8
o
c:::
0-
C
~
01 1010
c:::
:;
...
Z
-...
E
.0
::J
E
.2
LL
'0
o
G)
Zl~~--------~----------~
102
103
LP Problem Size (n)
(a)
104
~105 ~--------~---------~
102
103
10?
LP Problem Size (n)
(b)
Figure 1: Convergence rates as a function of n
applied to NN training in a way which preserves the computational efficiency of IP
solvers. The AM algorithm can be used to solve off-line problems such as codebook
generation and parameter identification in colony control applications. The method
is currently being used to solve hybrid control problems of the type in [9]. Areas of
future research concern the study of large-step IP methods and extensions of AM
training to other EM algorithms.
References
[1] S. Nowlan, "Maximum likelihood competitive learning," in Advances in Neural Information Processing Systems 2, pp. 574-582, San Mateo, California: Morgan Kaufmann
Publishers, Inc., 1990.
[2] M. Jordan and R. Jacobs, "Hierarchical mixtures of experts and the EM algorithm,"
Tech. Rep. 9301, MIT Computational Cognitive Science, Apr. 1993.
[3] R. Jacobs, M. Jordan, S. Nowlan, and G. Hinton, "Adaptive mixtures oflocal experts,"
Neural Computation, vol. 3, pp. 79-87, 1991.
[4] W. Byrne, "Alternating minimization and Boltzmann machine learning," IEEE Transactions on Neural Networks , vol. 3, pp. 612-620, July 1992.
[5] M. Jordan and 1. Xu, "Convergence results for the EM approach to mixtures of experts
architectures," Tech. Rep. 9303, MIT Computational Cognitive Science, Sept. 1993.
[6] C. Gonzaga, "Path-following methods for linear programming," SIAM Review, vol. 34,
pp. 167-224, June 1992.
[7] P. Szymanski and M. Lemmon, "A modified interior point method for supervisory controller design," in Proceedings of the 99rd IEEE Conference on Decision and Control,
pp. 1381-1386, Dec. 1994.
[8] M. Vidyasagar, Nonlinear Systems Analysis. Englewood Cliffs, New Jersey: PrenticeHall, Inc., 1993.
[9] M. Lemmon, J. Stiver, and P. Antsaklis, "Event identification and intelligent hybrid
control," in Hybrid Systems (R. L. Grossman, A. Nerode, A. P. Ravn, and H. Rischel,
eds.), vol. 736 of Lecture Notes in Computer Science, pp. 265-296, Springer-Verlag,
1993.
| 980 |@word inversion:1 nd:2 km:1 simulation:2 simplifying:1 contraction:1 jacob:2 thereby:1 solid:1 series:2 current:4 z2:2 nowlan:2 assigning:1 must:3 j1:3 update:28 pursued:2 parameterization:3 steepest:1 lr:4 nearness:1 provides:1 codebook:1 successive:8 consists:1 manner:4 examine:1 decreasing:2 solver:4 increasing:2 lib:1 notation:1 mass:1 ch1:2 cm:1 minimizes:1 developed:1 guarantee:3 nf:1 exactly:1 classifier:2 control:6 zl:1 ly:2 engineering:2 local:2 aat:1 cliff:1 path:10 approximately:2 might:1 nz:1 mateo:1 supq:1 mmlmlze:3 procedure:6 area:1 oflocal:1 convenient:1 word:1 radial:3 interior:8 close:3 pzq:2 conventional:1 go:1 convex:1 examines:1 stability:2 controlling:3 zzt:3 programming:3 pa:1 element:1 iib:2 electrical:2 solved:1 region:4 ensures:1 convexity:1 complexity:1 solving:5 upon:1 efficiency:4 basis:3 triangle:1 jersey:1 ilh:1 distinct:1 fast:1 describe:1 neighborhood:1 quite:1 solve:3 ip:10 sequence:10 inserting:1 ent:1 convergence:26 chl:3 produce:3 guaranteeing:1 rotated:1 colony:1 predicted:3 involves:1 implies:3 convention:1 closely:1 escent:1 stochastic:2 viewing:1 proposition:1 extension:1 sufficiently:1 normal:1 exp:1 mapping:2 tor:1 cck:1 estimation:1 currently:2 largest:1 successfully:1 establishes:1 minimization:15 mit:2 modified:2 ck:10 cr:1 derived:1 ax:4 june:1 notational:1 rank:2 likelihood:1 tech:2 am:17 stopping:2 nn:1 mth:12 irn:9 arg:1 dual:1 denoted:2 development:1 proposes:1 constrained:2 initialize:1 having:1 future:1 minimized:1 logx:1 intelligent:1 randomly:1 preserve:3 resulted:2 floating:1 englewood:1 introduces:1 mixture:3 notre:4 nl:3 primal:6 necessary:2 iieii:1 desired:1 theoretical:2 soft:2 maximization:1 cost:6 subset:1 too:2 perturbed:3 combined:1 density:1 siam:1 off:1 regressor:6 michael:5 ym:5 prenticehall:1 central:15 choose:1 cognitive:2 expert:4 grossman:1 aggressive:2 inc:2 depends:1 later:2 break:1 fol:2 wm:3 competitive:2 maintains:2 contribution:1 om:2 square:6 ir:4 accuracy:1 kaufmann:1 characteristic:1 efficiently:1 identification:2 simultaneous:1 fo:1 ed:1 pp:6 dm:2 associated:2 proof:2 supervised:1 specify:1 response:2 formulation:1 done:1 until:2 receives:1 nonlinear:2 supervisory:1 effect:1 concept:1 consisted:1 counterpart:2 byrne:1 aggressively:2 alternating:11 ll:2 sin:4 noted:1 criterion:3 hilus:1 ranging:1 recently:1 functional:3 ji:1 extend:1 significant:1 enjoyed:1 rd:1 ssd:4 f0:1 longer:1 add:1 recent:3 apart:1 verlag:1 inequality:1 rep:2 iih211:4 yi:8 morgan:1 minimum:1 preceding:5 mr:3 converge:9 july:1 ii:9 full:2 characterized:2 plugging:1 controlled:1 prediction:2 ile:1 controller:1 expectation:1 metric:1 iteration:9 represent:1 dec:1 receive:1 preserved:1 separately:1 decreased:1 addressed:1 publisher:1 zw:1 subject:3 ofx:1 jordan:3 ee:2 iii:1 zi:14 architecture:1 reduce:1 peter:5 specifies:1 exist:1 arising:1 zd:1 discrete:1 vol:4 key:1 prevent:1 monotone:5 year:1 run:2 parameterized:2 vn:1 decision:1 scaling:1 bit:1 bound:5 ct:1 dame:4 guaranteed:2 hi:1 quadratic:6 argument:1 min:3 alternate:1 nerode:1 combination:1 em:7 unity:2 wi:1 lp:26 modification:4 hl:14 maxy:1 leg:4 xo:1 k2l:2 remains:1 locus:1 lyt:1 adopted:1 operation:1 hierarchical:1 away:1 appropriate:1 original:1 denotes:1 remaining:1 ensure:2 approximating:1 said:1 exhibit:2 kth:4 distance:1 assuming:2 ru:1 length:1 minimizing:4 statement:1 ilw:1 stated:1 implementation:4 design:1 zt:2 boltzmann:2 allowing:2 neuron:16 observation:1 finite:3 lemmon:8 descent:1 flop:2 hinton:1 community:1 introduced:1 pair:2 required:1 connection:1 california:1 established:2 ctx:2 program:1 max:3 vidyasagar:1 event:1 force:1 hybrid:3 zr:2 sept:1 review:1 multiplication:1 embedded:1 lecture:1 highlight:1 sublinear:4 generation:1 proven:1 approximator:3 h2:1 agent:6 xp:1 repeat:1 lier:1 allow:1 taking:1 barrier:4 sparse:1 boundary:1 dimension:1 llc:2 valid:3 computes:2 adaptive:1 regressors:4 san:1 lz:3 far:2 transaction:1 approximate:1 keep:1 reveals:1 xi:2 szymanski:6 search:2 nature:1 investigated:1 diag:1 apr:1 arise:1 xu:1 augmented:4 referred:3 xl:1 lie:5 breaking:1 weighting:1 learns:1 theorem:3 specific:1 gating:9 er:3 normalizing:1 closeness:2 concern:1 conditioned:1 gap:4 logarithmic:1 expressed:2 springer:1 corresponds:2 minimizer:2 relies:1 conditional:2 viewed:4 kee:1 consequently:1 rbf:3 considerable:1 hard:1 feasible:1 change:1 principal:1 called:2 total:2 duality:4 newtonraphson:1 dept:2 d1:2 |
7,222 | 981 | Reinforcement Learning with Soft State
Aggregation
Satinder P. Singh
singh@psyche.mit.edu
Tommi Jaakkola
tommi@psyche.mit.edu
Michael I. Jordan
jordan@psyche.mit.edu
Dept. of Brain & Cognitive Sciences (E-lO)
M.I.T.
Cambridge, MA 02139
Abstract
It is widely accepted that the use of more compact representations
than lookup tables is crucial to scaling reinforcement learning (RL)
algorithms to real-world problems. Unfortunately almost all of the
theory of reinforcement learning assumes lookup table representations. In this paper we address the pressing issue of combining
function approximation and RL, and present 1) a function approximator based on a simple extension to state aggregation (a commonly used form of compact representation), namely soft state
aggregation, 2) a theory of convergence for RL with arbitrary, but
fixed, soft state aggregation, 3) a novel intuitive understanding of
the effect of state aggregation on online RL, and 4) a new heuristic
adaptive state aggregation algorithm that finds improved compact
representations by exploiting the non-discrete nature of soft state
aggregation. Preliminary empirical results are also presented.
1
INTRODUCTION
The strong theory of convergence available for reinforcement learning algorithms
(e.g., Dayan & Sejnowski, 1994; Watkins & Dayan, 1992; Jaakkola, Jordan & Singh,
1994; Tsitsiklis, 1994) makes them attractive as a basis for building learning control architectures to solve a wide variety of search, planning, and control problems.
Unfortunately, almost all of the convergence results assume lookup table representa-
362
Satinder Singh, Tommi Jaakkola, Michael I. Jordan
tions for value functions (see Sutton, 1988; Dayan, 1992; Bradtke, 1993; and Vanroy
& Tsitsiklis, personal communication; for exceptions). It is widely accepted that
the use of more compact representations than lookup tables is crucial to scaling RL
algorithms to real-world problems.
In this paper we address the pressing issue of combining function approximation and
RL, and present 1) a function approximator based on a simple extension to state
aggregation (a commonly used form of compact representation, e.g., Moore, 1991),
namely soft state aggregation, 2) a theory of convergence for RL with arbitrary, but
fixed, soft state aggregation, 3) a novel intuitive understanding of the effect of state
aggregation on online RL, and 4) a new heuristic adaptive state aggregation algorithm that finds improved compact representations by exploiting the non-discrete
nature of soft state aggregations. Preliminary empirical results are also presented.
Problem Definition and Notation: We consider the problem of solving large
Markovian decision processes (MDPs) using RL algorithms and compact function
approximation. We use the following notation: 8 for state space, A for action space,
pa(s, s') for transition probability, Ra(s) for payoff, and 'Y for discount factor. The
objective is to maximize the expected, infinite horizon, discounted sum of payoffs.
1.1
FUNCTION APPROXIMATION: SOFT STATE CLUSTERS
In this section we describe a new function approximator (FA) for RL. In section 3
we will analyze it theoretically and present convergence results. The FA maps the
state space 8 into M > 0 aggregates or clusters from cluster space X. Typically,
M < < 181. We allow soft clustering, where each state s belongs to cluster x with
probability P(xls), called the clustering probabilities. This allows each state s to
belong to several clusters. An interesting special case is that of the usual state
aggregation where each state belongs only to one cluster. The theoretical model is
that the agent can observe the underlying state but can only update a value function
for the clusters. The value of a cluster generalizes to all states in proportion to
the clustering probabilities. Throughout we use the symbols x and y to represent
individual clusters and the symbols sand s' to represent individual states.
2
A GENERAL CONVERGENCE THEOREM
An online RL algorithm essentially sees a sequence of quadruples, < St, at, StH, rt >,
representing a transition from current state St to next state StH on current action
at with an associated payoff rt. We will first prove a general convergence theorem
for Q-learning (Watkins & Dayan, 1992) applied to a sequence of quadruples that
mayor may not be generated by a Markov process (Bertsekas, 1987). This is
required because the RL problem at the level of the clusters may be non-Markovian.
Conceptually, the sequence of quadruples can be thought of as being produced by
some process that is allowed to modify the sequence of quadruples produced by a
Markov process, e.g., by mapping states to clusters. In Section 3 we will specialize
the following theorem to provide specific results for our function approximator.
Consider any stochastic process that generates a sequence of random quadruples,
\]! = {< Xi, ai, Yi, ri > h, where Xi, Yi E Y, ai E A, and ri is a bounded real number.
Note that Xi+l does not have to be equal to Yi. Let IYI and IAI be finite, and define
Reinforcement Learning with Soft State Aggregation
363
indicator variables
Xi ( x, a, Y) = {
and
Wi =< x,a,y,. > (for any r)
?1 owhen
th erwlse,
.
when Wi =< x, a, .,.
otherwise.
> (for any
y, and any r)
Define
,a ,( ) -_ 2:t-iXk(x,a,y)
PIJx,y
,
,
2:1=i n(x, a)
and
Theorem 1: If V( > 0, 3ME < 00, such that for all i ~ 0, for all x, y E Y, and
for all a E A, the following conditions characterize the infinite sequence W: with
probability 1 - (,
IPi~i+M.(X,y)-pa(x,y)1
< (
IRi,i+MJX) - Ra(x)1 < (,
and
(1)
where for all x, a, and y, with probability one Pg,oo(x, y) = pa(x, y), and Rg,oo(x) =
Ra(x). Then, online Q-learning applied to such a sequence will converge with
probability one to the solution of the following system of equations: Vx E Y, and
Va E A,
Q(x, a) = Ra(x) + 'Y " pa(x, y) maxQ(y, a')
L.J
a'EA
(2)
yEY
Proof: Consider the semi-batch version of Q-learning that collects the changes
to the value function for M steps before making the change. By assumption, for
any (, making ME large enough will ensure that with probability 1 - (, the sample
(x, y) and Ri,i+M
quantities for the ith batch, Pti+M
,?.)
< (i)
? (x) are within ( of the
asymptotic quantities. In Appendix A we prove that the semi-batch version of Qlearning outlined above converges to the solution of Equation 2 with probability one.
The semi-batch proof can be extended to online Q-learning by using the analysis
developed in Theorem 3 of Jaakkola et al. (1994). In brief, it can be shown that
the difference caused by the online updating vanishes in the limit thereby forcing
semi-batch Q-learning and online Q-learning to be equal asymptotically. The use
of the analysis in Theorem 3 from Jaakkola et al. (1994) requires that the learning
rate parameters CY are such that maxlEM.(Ic)al
alex)
() _
1 uniformly w.p.l.; ME(k) is
x
the kth batch of size ME' If CYt(x) is non-increasing in addition to satisfying the
conventional Q-learning conditions, then it will also meet the above requirement .
o
Theorem 1 provides the most general convergence result available for Q-learning
(and TD(O?; it shows that for an arbitrary quadruple sequence satisfying the ergodicity conditions given in Equations 1, Q-learning will converge to the solution
of the MDP constructed with the limiting probabilities (Po,oo) and payoffs (Ro,oo).
Theorem 1 combines and generalizes the results on hard state aggregation and
value iteration presented in Vanroy & Tsitsiklis (personal communication), and on
partially observable MDPs in Singh et al. (1994).
364
3
Satinder Singh, Tommi Jaakkola, Michael I. Jordan
RL AND SOFT STATE AGGREGATION
In this section we apply Theorem 1 to provide convergence results for two cases: 1)
using Q-Iearning and our FA to solve MDPs, and 2) using Sutton's (1988) TD(O)
and our FA to determine the value function for a fixed policy. As is usual in online
RL, we continue to assume that the transition probabilities and the payoff function
of the MDP are unknown to the learning agent. Furthermore, being online such
algorithms cannot sample states in arbitrary order. In this section, the clustering
probabilities P(xls) are assumed to be fixed.
Case 1: Q-learning and Fixed Soft State Aggregation
Because of function approximation, the domain of the learned Q-value function is
constrained to be X x A (X is cluster space). This section develops a "Bellman
equation" (e.g., Bertsekas, 1987) for Q-Iearning at the level of the cluster space. We
assume that the agent follows a stationary stochastic policy 7r that assigns to each
state a non-zero probability of executing every action in every state. Furthermore,
we assume that the Markov chain under policy 7r is ergodic. Such a policy 7r is a
persistently exciting policy. Under the above conditions p'II"(slx) = :,
;C;,:;p$"C$')'
where for all s, p'll"(s) is the steady-state probability of being in state s.
Corollary 1: Q-Iearning with soft state aggregation applied to an MDP while
following a persistently exciting policy 7r will converge with probability one to the
solution of the following system of equations: Vex, a) E (X x A),
E$'
and pa(s, y) =
pa(s, s')P(yls'). The Q-value function for the state space can
then be constructed via Q(s, a) = Ex P(xls)Q(x, a) for all (s, a).
Proof: It can be shown that the sequence of quadruples produced by following policy 7r and independently mapping the current state s to a cluster x with probability
P(xls) satisfies the conditions of Theorem 1. Also, it can be shown that
Note that the Q-values found by clustering are dependent on the sampling policy
7r, unlike the lookup table case.
Case 2: TD(O) and Fixed Soft State Aggregation
We present separate results for TD(O) because it forms the basis for policy-iterationlike methods for solving Markov control problems (e.g., Barto, Sutton & Anderson,
1983) - a fact that we will use in the next section to derive adaptive state aggregation methods. As before, because of function approximation, the domain of the
learned value function is constrained to be the cluster space X.
Corollary 2: TD(O) with soft state aggregation applied to an MDP while following
a policy 7r will converge with probability one to the solution of the following system
Reinforcement Learning with Soft State Aggregation
365
of equations: Vx EX,
V(x)
(4)
where again as in Q-Iearning the value function for the state space can be constructed via V(s) = Lx P(xls)V(x) for all s.
Proof: Corollary 1 implies Corollary 2 because TD(O) is a special case of Q-Iearning
for MDPs with a single (possibly randomized) action in each state. Equation 4
provides a "Bellman equation" for TD(O) at the level of the cluster space. 0
4
ADAPTIVE STATE AGGREGATION
In previous sections we restricted attention to a function approximator that had a
fixed compact representation. How might one adapt the compact representation online in order to get better approximations of value functions? This section presents
a novel heuristic adaptive algorithm that improves the compact representation by
finding good clustering probabilities given an a priori fixed number of clusters. Note
that for arbitrary clustering, while Corollaries 1 and 2 show that RL will find solutions with zero Bellman error in cluster space, the associated Bellman error in
the state space will not be zero in general. Good clustering is therefore naturally
defined in terms of reducing the Bellman error for the states of the MDP.
Let the clustering probabilities be parametrized as follows P(xls; 0)
= L:~~:';~/ . ),
where o( x , s) is the weight between state s and cluster x. Then the Bellman error
at .state s given parameter (a matrix) is,
?
J(sIO)
V( slO) - [R' (s)
+ r ~ P' (8, s')V( S'IO)]
~ P(xls; O)V(xIO) -
[w(s)
+ r ~ P'(s, s') ~ P(xls'; O)V(XIO)]
Adaptive State Aggregation (ASA) Algorithm:
Step 1: Compute V(xIO) for all x E X using the TD(O) algorithm.
2 (9)
Step 2.. Let !:l.0 -- -a oJ89
. Go to step 1.
where Step 2 tries to minimize the Bellman error for the states by holding the
cluster values fixed to those computed in Step 1. We have
8J 2 (sI0)
80(y, s)
2J(sI0) [P(yls; 0)(1 - ,p7r(s, s?(V(yIO) - V(sIO?].
The Bellman error J(sIO) cannot be computed directly because the transition probabilities P( s, s') are unknown. However, it can be estimated by averaging the sample
366
Sa tinder Singh, Tommi Jaakkola, Michael I. Jordan
Bellman error. P(yls; B) is known, and (1 - ,P 7r (s, s)) is always positive, and independent of y, and can therefore be absorbed into the step-size c?. The quantities
V(yIB) and V(sIB) are available at the end of Step 1. In practice, Step 1 is only
carried out partially before Step 2 is implemented. Partial evaluation works well
because the changes in the clustering probabilities at Step 2 are small, and because
the final V(xIB) at the previous Step 1 is used to initialize the computation of V(xIB)
at the next Step 1.
------ 2 Clusters
--------- 4 Clusters
to Clusters
- - 20 Clusters
4
..
,' .
.. ",' " ....... - -.. .
"
. ... " .... .' . . " .
Iterations of ASA
Figure 1: Adaptive State Clustering. See text for explanation.
Figure 1 presents preliminary empirical results for the ASA algorithm. It plots the
squared Bellman error summed over the state space as a function of the number of
iterations of the ASA algorithm with constant step-size c?. It shows error curves
for 2, 4, 10 and 20 clusters averaged over ten runs of randomly constructed 20
state Markov chains. Figure 4 shows that ASA is able to adapt the clustering
probabilities to reduce the Bellman error in state space, and as expected the more
clusters the smaller the asymptotic Bellman error. In future work we plan to test
the policy iteration version of the adaptive soft aggregation algorithm on Markov
control problems.
5
SUMMARY AND FUTURE WORK
Doing RL on aggregated states is potentially very advantageous because the value of
each cluster generalizes across all states in proportion to the clustering probabilities.
The same generalization is also potentially perilous because it can interfere with the
contraction-based convergence of RL algorithms (see Vee, 1992; for a discussion).
This paper resolves this debate for the case of soft state aggregation by defining a set
of Bellman Equations (3 and 4) for the control and policy evaluation problems in the
non-Markovian cluster space, and by proving that Q-Iearning and TD(O) solve them
respectively with probability one. Theorem 1 presents a general convergence result
that was applied to state aggregation in this paper, but is also a generalization of
the results on hidden state presented in Singh et al. (1994), and may be applicable
Reinforcement Learning with Soft State Aggregation
367
to other novel problems. It supports the intuitive picture that if a non-Markovian
sequence of state transitions and payoffs is ergodic in the sense of Equation 1, then
RL algorithms will converge w.p.l. to the solution of an MDP constructed with the
limiting transition probabilities and payoffs.
We also presented a new algorithm, ASA, for adapting compact representations,
that takes advantage of the soft state aggregation proposed here to do gradient descent in clustering probability space to minimize squared Bellman error in the state
space. We demonstrated on simple examples that ASA is able to adapt the clustering probabilities to dramatically reduce the Bellman error in state space. In future
work we plan to extend the convergence theory presented here to discretizations of
continuous state MDPs, and to further test the ASA algorithms.
A
Convergence of semi-batch Q-Iearning (Theorem 1)
Consider a semi-batch algorithm that collects the changes to the Q-value function
for M steps before making the change to the Q-value function. Let
kM
R%(x) =
L
kM
L
Mk(X, a) =
riXi(x, a);
i=(k-l)M
Xi(X, a)
i=(k-l)M
and
kM
Mk(x,a,y)
L
=
Xi(x,a,y)
i=(k-l)M
Then the Q-value of (x, a) after the
Qk+l(X, a)
=
kth
batch is given by:
(1- Mk(X, a)ak(x, a))Qk(x, a)
Rk(x)
+Mk(x, a) ak(x, a) [ M kx,a
( )
Let
Q be
Mk(X, a, y)
+, ~
~ M (
kx,a ) m;pcQk(Y, a)
I
]
a
yEY
the solution to Equation 2. Define,
Fk(x, a)
=
Rk(x)
M (
)
kx,a
~
+, ~
yEY
Mk(x, a, y)
,M ( ) m~xQk(Y, a) - Q(x, a),
kx,a
a
then, if Vk(X) = maxa Qk(X, a) and Vex) = maxa Q(x, a),
Fk(x, a) =
~
Mk(X, a, y)
) (Vk(y) - V(y)]
kx,a
y
,~ M (
[(Mk(x,a,y)
+, ~
~
Mk(X,a)
a
(
Rk(x)
a
+ (Mkx,a
( ) - Ro,oo(x))
))-()]
- PO,oo x,y V y
,
y
The quantity Fk(x, a) can be bounded by
IIFk(x, a) II
:s; ,llVk - VII + II(~~~~L
- Rg,oo(x))11
+,11 Ly(~tX~~~y - Po,oo(x, y))V(y)11 :s; ,llVk - VII + Cfr,
a
M? th e Iarger 0 f IMk(x,a)
R&(x) - R 0,00
()I
I ~ (Mk(x,a,y)
na ( x, y ))1 . By
h
were
fk 1S
x , an d 'l..Jy
Mk(x,a)
.ro,oo
368
Sa tinder Singh, Tommi Jaakkola, Michael I. Jordan
assumption for any (. > 0, 3Mf < 00 such that (.~< < (. with probability 1 - c The
variance of Fk(X, a) can also be shown to be bounded because the variance of the
sample probabilities is bounded (everything else is similar to standard Q-learning
for MDPs). Therefore by Theorem 1 of Jaakkola et al. (1994), f<2r any (. > 0,
with probability (1 - f), Qk(X, a) - Qoo(X, a), where IQoo(x, a) - Q(x, a)1 ~ Cc
Therefore, semi-batch Q-learning converges with probability one. 0
Acknowledgements
This project was supported in part by a grant from the McDonnell-Pew Foundation,
by a grant from ATR Human Information Processing Research Laboratories, and by
a grant from Siemens Corporation. Michael I. Jordan is a NSF Presidential Young
Investigator.
References
A. G. Barto, R. S. Sutton, & C. W. Anderson. (1983) Neuronlike elements that can
solve difficult learning control problems. IEEE SMC, 13:835-846.
D. P. Bertsekas. (1987) Dynamic Programming: Deterministic and Stochastic Models, Prentice-Hall.
S. J. Bradtke. (1993) Reinforcement learning applied to linear quadratic regulation.
In Advances in Neural Information Processing Systems 5, pages 295-302.
P. Dayan. (1992) The convergence of TD(A) for general A. Machine Learning,
8(3/4):341-362.
P. Dayan & T.J. Sejnowski. (1994) TD(A) converges with probability 1. Machine
Learning, 13(3).
T. Jaakkola, M. I. Jordan, & S. P. Singh. (1994) On the convergence of stochastic
iterative dynamic programming algorithms. Neural Computation, 6(6):11851201.
A. W. Moore. (1991) Variable resolution dynamic programming: Efficiently learning
action maps in multivariate real-valued state-spaces. In M aching Learning:
Proceedings of the Eighth International Workshop, pages 333- 337.
S. P. Singh, T. Jaakkola, & M. I. Jordan. (1994) Learning without state-estimation
in partially observable markovian decision processes. In Machine Learning:
Proceedings of the Eleventh International Conference, pages 284-292.
R. S. Sutton. (1988) Learning to predict by the methods of temporal differences.
Machine Learning, 3:9-44.
J. Tsitsiklis. (1994) Asynchronous stochastic approximation and Q-learning. Machine Learning, 16(3):185-202.
B. Vanroy & J. Tsitsiklis. (personal communication)
C. J. C. H. Watkins & P. Dayan. (1992) Q-learning. Machine Learning, 8(3/4):279292.
R. C. Yee. (1992) Abstraction in control learning. Technical Report COINS Technical Report 92-16, Department of Computer and Information Science, University
of Massachusetts, Amherst, MA 01003. A dissertation proposal.
| 981 |@word version:3 proportion:2 advantageous:1 km:3 contraction:1 pg:1 thereby:1 current:3 plot:1 update:1 stationary:1 ith:1 dissertation:1 provides:2 lx:1 ipi:1 constructed:5 prove:2 specialize:1 combine:1 eleventh:1 theoretically:1 expected:2 ra:4 planning:1 brain:1 bellman:15 discounted:1 td:11 resolve:1 increasing:1 project:1 notation:2 underlying:1 bounded:4 maxa:2 developed:1 finding:1 corporation:1 temporal:1 every:2 iearning:7 ro:3 control:7 ly:1 grant:3 bertsekas:3 before:4 positive:1 modify:1 limit:1 io:1 sutton:5 ak:2 quadruple:7 meet:1 might:1 collect:2 smc:1 slx:1 averaged:1 practice:1 discretizations:1 empirical:3 thought:1 adapting:1 yio:1 get:1 cannot:2 prentice:1 yee:1 conventional:1 map:2 demonstrated:1 deterministic:1 go:1 iri:1 attention:1 independently:1 ergodic:2 resolution:1 assigns:1 proving:1 limiting:2 programming:3 pa:6 persistently:2 element:1 satisfying:2 updating:1 cy:1 vanishes:1 xio:3 dynamic:3 personal:3 singh:11 solving:2 asa:8 basis:2 po:3 tx:1 describe:1 sejnowski:2 aggregate:1 heuristic:3 widely:2 solve:4 valued:1 otherwise:1 presidential:1 final:1 online:10 sequence:10 pressing:2 advantage:1 combining:2 intuitive:3 exploiting:2 convergence:15 cluster:28 requirement:1 converges:3 executing:1 tions:1 oo:9 derive:1 sa:2 strong:1 implemented:1 implies:1 tommi:6 stochastic:5 vx:2 human:1 everything:1 sand:1 generalization:2 preliminary:3 extension:2 hall:1 ic:1 mapping:2 predict:1 estimation:1 applicable:1 si0:2 mit:3 cyt:1 always:1 representa:1 qoo:1 barto:2 jaakkola:11 corollary:5 vk:2 sense:1 dayan:7 dependent:1 abstraction:1 typically:1 hidden:1 issue:2 priori:1 plan:2 constrained:2 special:2 initialize:1 summed:1 equal:2 sampling:1 future:3 report:2 develops:1 randomly:1 individual:2 neuronlike:1 evaluation:2 chain:2 partial:1 theoretical:1 mk:11 soft:20 markovian:5 characterize:1 st:2 international:2 randomized:1 amherst:1 michael:6 na:1 again:1 squared:2 possibly:1 cognitive:1 lookup:5 caused:1 try:1 ixk:1 analyze:1 doing:1 aggregation:30 minimize:2 qk:4 variance:2 efficiently:1 conceptually:1 produced:3 cc:1 tinder:2 definition:1 naturally:1 associated:2 proof:4 massachusetts:1 improves:1 ea:1 improved:2 iai:1 anderson:2 furthermore:2 ergodicity:1 interfere:1 mdp:6 building:1 effect:2 moore:2 laboratory:1 attractive:1 ll:1 steady:1 slo:1 bradtke:2 novel:4 rl:18 belong:1 extend:1 cambridge:1 ai:2 pew:1 outlined:1 fk:5 had:1 xib:2 iyi:1 multivariate:1 belongs:2 forcing:1 yls:3 continue:1 yi:3 p7r:1 converge:5 maximize:1 determine:1 aggregated:1 semi:7 ii:3 sib:1 technical:2 adapt:3 jy:1 va:1 essentially:1 mayor:1 iteration:4 represent:2 proposal:1 addition:1 else:1 crucial:2 unlike:1 jordan:10 enough:1 variety:1 architecture:1 reduce:2 action:5 dramatically:1 discount:1 ten:1 nsf:1 estimated:1 discrete:2 asymptotically:1 sum:1 run:1 almost:2 throughout:1 vex:2 decision:2 appendix:1 scaling:2 quadratic:1 alex:1 ri:3 generates:1 department:1 mcdonnell:1 smaller:1 across:1 pti:1 psyche:3 sth:2 wi:2 making:3 restricted:1 equation:11 end:1 available:3 generalizes:3 apply:1 observe:1 batch:10 coin:1 assumes:1 clustering:15 ensure:1 objective:1 quantity:4 fa:4 rt:2 usual:2 gradient:1 kth:2 separate:1 atr:1 parametrized:1 me:4 cfr:1 vee:1 difficult:1 unfortunately:2 regulation:1 potentially:2 holding:1 debate:1 policy:12 unknown:2 rixi:1 markov:6 finite:1 descent:1 payoff:7 extended:1 communication:3 defining:1 arbitrary:5 namely:2 required:1 yey:3 learned:2 maxq:1 address:2 able:2 eighth:1 explanation:1 indicator:1 representing:1 mdps:6 brief:1 picture:1 carried:1 text:1 understanding:2 acknowledgement:1 asymptotic:2 interesting:1 approximator:5 foundation:1 agent:3 exciting:2 lo:1 summary:1 supported:1 asynchronous:1 tsitsiklis:5 allow:1 wide:1 curve:1 world:2 transition:6 commonly:2 reinforcement:8 adaptive:8 xqk:1 compact:11 qlearning:1 observable:2 satinder:3 assumed:1 xi:6 search:1 continuous:1 iterative:1 table:5 nature:2 erwlse:1 domain:2 allowed:1 yib:1 xl:8 watkins:3 young:1 theorem:13 rk:3 specific:1 symbol:2 workshop:1 horizon:1 kx:5 mf:1 vii:2 rg:2 absorbed:1 partially:3 satisfies:1 ma:2 change:5 hard:1 infinite:2 uniformly:1 reducing:1 averaging:1 called:1 accepted:2 siemens:1 exception:1 sio:3 support:1 investigator:1 dept:1 ex:2 |
7,223 | 982 | Coarse-to-Fine Image Search Using Neural
Networks
Clay D. Spence, John C. Pearson, and Jim Bergen
Nationallnfonnation Display Laboratory
P.O. Box 8619
Princeton, NJ 08543-8619
cspence@sarnoff.com
John_Pearson@maca.sarnoff.com
jbergen@sarnoff.com
Abstract
The efficiency of image search can be greatly improved by using a
coarse-to-fine search strategy with a multi-resolution image representation. However, if the resolution is so low that the objects have few distinguishing features, search becomes difficult. We show that the
performance of search at such low resolutions can be improved by using
context information, i.e., objects visible at low-resolution which are not
the objects of interest but are associated with them. The networks can be
given explicit context information as inputs, or they can learn to detect
the context objects, in which case the user does not have to be aware of
their existence. We also use Integrated Feature Pyramids, which represent high-frequency information at low resolutions. The use of multiresolution search techniques allows us to combine information about the
appearance of the objects on many scales in an efficient way. A natural
fOlm of exemplar selection also arises from these techniques. We illustrate these ideas by training hierarchical systems of neural networks to
find clusters of buildings in aerial photographs of farmland.
1
INTRODUCTION
Coarse-to-fine image search is a general technique for improving the efficiency of any
search algorithm. (Burt, 1988a and b; Burt et al., 1989). One begins by searching a lowresolution (coarse-scale) version of the image, typically obtained by constructing an image
982
Clay D. Spence, John C. Pearson, Jim Bergen
pyramid (Burt and Adelson, 1983). Since this version is smaller than the original, there are
far fewer pixels to be processed and the search is correspondingly faster. To improve the
certainty of detection and refine the location estimates, the search process is repeated at
higher resolution (finer scale), but only in those regions-of-interest (ROls) which were
identified at lower resolution as likely to contain one of the objects to be found, thus
greatly reducing the actual area to be searched. This process is repeated at successively
higher resolutions until sufficient certainty and precision is achieved or the original image
has been searched. 1
These pyramid techniques scale well with image size and can quickly process large
images, but the relatively simple, hand-tuned pattern recognition components that are typically used limit their accuracy. Neural networks provide a complementary set of techniques, since they can learn complex patterns, but they don't scale well with image size.
We have developed a Hybrid PyramidlNeural Network (HPNN) system that combines the
two techniques so as to leverage their strengths and compensate for their weaknesses.
A novel benefit of combining pyramids and neural networks is the potential for automatically learning to use context information, by which we mean any visible characteristics of
the object's surroundings which tend to distinguish it from other objects. Examples of
context information which might be useful in searching aerial photographs for man-made
objects are the proximity of roads and terrain type. Pyramids provide the ability to detect
context in a low-resolution image representation, which is necessary for efficient processing of large regions of the image. Neural networks provide the ability to discover context
of which we may be unaware, and to learn the relevance of the context to detection, the
dependence of detection probability on distance from a context object, etc. Context can
help to narrow the search region at low resolutions and improve the detection performance
since it is, by definition, relevant information. Of course, the usefulness of the idea will be
different for each problem.
Context can be explicitly provided, e.g., in the form of a road map. As mentioned above, a
neural net may also learn to exploit some context which is not explicitly provided, but
must be inferred from the image data. If there is such context and the network's architecture and input features are capable of exploiting it, it should learn to do so even if we use
ordinary training methods that do not explicitly take context into account.
We currently implement these functions in a hierarchical sequence of ordinary feed-forward networks with sigmoidal units, one network at each resolution in an image pyramid
(Fig. 1). Each network receives some features extracted from a window in the image, and
its output is interpreted as the probability of finding one of the searched-for objects in the
center of the window. The network is scanned over all ROIs in this manner. (At the lowest
resolution there is one ROI, which is the entire image.)
The networks are trained one at a time, starting at the lowest resolution. Context is made
available to networks at higher resolutions by giving them information from the lower-resolution networks as additional inputs, which we will call context inputs. These could be
I Several authors have investigated multi-scale processing of one-dimensional signals with neural
networks, e.g., Mozer (1994) and Burr and Miyata (1993) studied music composition. Burr and
Miyata use sub-sampling as in a pyramid decomposition. Images differ somewhat from music in that
there are primitive features at all scales (limited by the sampling rate and image size), whereas the
average frequency spectrum of music over a long time span doesn't seem likely to be meaningful.
The original paper on pyramid image representation is (Burt and Adelson, 1983).
Coarse-To-Fine Image Search Using Neural Networks
983
Target
High-Resolution Features
Probability
Context Information
Control of Search Region
Features
Figure 1. A Hierarchical Search System that Exploits Context.
taken from either the hidden units or output units of the lower-resolution networks. By
training the networks in sequence starting at low resolution, we can choose to train the
higher-resolution networks in the ROls selected by the lower-resolution networks, thus
providing a simple means of exemplar selection. This coarse-to-fine training is often useful, since many problems have relatively few positive examples per image, but because of
the size of the image there are an enormous number of more-or-Iess redundant negative
examples.
2
ANEXAMPLEPROBLEM
To demonstrate these ideas we applied them to the problem of finding clusters of buildings
in aerial photographs of a rural area. Since buildings are almost always near a road, roads
are context objects associated with buildings. We approached the problem in two ways to
demonstrate different capabilities. First, we trained systems of networks with explicitl~
provided context in the form of road maps, in order to demonstrate the possible benefits.
Second, we trained systems without explicitly-provided context, in order to show that the
context could be learned.
2.1
EXPLICIT CONTEXT
For comparison purposes, we trained one network to search a high-resolution image
directly, with no explicit context inputs and no inputs from other networks. To demonstrate the effect of learned context on search we trained a second system with no explicit
2 It would be surprising if this extra infonnation didn't help, since we know it is relevant. For
many applications digitized maps will be available, so demonstrating the possible perfonnance benefit is still worth-while.
984
Clay D. Spence, John C. Pearson, Jim Bergen
context provided to the nets, but each received inputs from all of the networks at lower
resolutions. These inputs were simply the outputs of those lower-resolution networks. To
demonstrate the benefit of explicitly provided context, we trained a third system with both
context inputs from lower-resolution networks and explicit context in the form of a road
map of the area, which was registered with the image.
2.1.1 Features
To preserve some distinguishing features at a given low resolution, we extracted simple
features at various resolutions and represented them at the lower resolution. The low-resolution representations were constructed by reducing the feature images with the usual
blur-and-sub-sample procedure used to construct a Gaussian pyramid (Burt and Adelson,
1983). The features should not be too computationally expensive to extract, otherwise the
efficiency benefit of coarse-to-fine search would be canceled.
The features used as inputs to the neural nets in the building-search systems were simple
measures of the spatial image energy in several frequency bands. We constructed these
feature images by building the Laplacian pyramid of the image3 and then taking the absolute values of the pixels in each image in the pyramid. We then constructed a Gaussian
pyramid of each of these images, to provide versions of them at different resolutions. A
neural net searching a given resolution received input from each energy image derived
from the same resolution and all higher resolutions. 4
Binary road-map images were constructed from the digitized aerial photographs. They
were reduced in resolution by first perfonning a binary blur of the image and then subsampling it by two in each dimension. In the binary blur procedure each pixel is set to one
if any of its nearest neighbors were road pixels before blurring. This is repeated to get road
maps at each resolution. To give the networks a rough measure of distance to a road, we
gave the nets inputs from linearly-blurred versions of the road maps, which were made by
expanding even lower-resolution versions of the road map with the same linear expand
operation used in constructing the Laplacian pyramid. These blurred road maps are therefore not binary. The networks at the fifth and third pyramid levels received inputs from the
road-maps at their resolution and from the two lower resolutions, while the network at the
first pyramid level received input from the road map at its resolution and the next lower
resolution.
2.1.2 Training the Networks
To estimate the probability of finding a building cluster at or around a given location in the
image, each network received a single pixel from the same location in all of its input
images. This should be adequate for search, since a single pixel in a feature image contains information about an extended region of the original. With the features we used, it
also makes the system invariant to rotations, so the networks do not have to learn this
invariance Also for simplicity, we did not train nets at all resolutions, but only at the fifth,
third and first pyramid levels, i.e., on images which were one-thirty-second, one-eighth,
3 The Laplacian pyramid is usually constructed by expanding the lower-resolution levels of a
Gaussian pyramid and subtracting each from the next-higher-resolution level. This gives a set of
images which are band-passed with one-octave spacing between the bands.
4 There are many examples of more sophisticated features, e.g., Lane, et aI., 1992, Ballard and
Wixson, 1993, and Greenspan, et aI., 1994. Simple features were adequate for demonstrating the
ideas of this paper.
Coarse-To-Fine Image Search Using Neural Networks
No Road Maps or context
985
100
,- -
Context
~
.....
.iii
o
--
-:'
.
~
I
90
r
,, ,
85
I
Q)
...
:::l
t-
~
- -
95
.~
a..
r ,'-
1-
80~~~~~~~~~~~~~~~~
a
5
15
10
% Folse Positives
20
Figure 2. ROC Curves for the Three Network Search Systems.
and one-half the width and height of the original image. A typical building cluster has a
linear extent of about 30 pixels in the original images, so in the fifth Gaussian pyramid
level they have few distinguishing features. The networks were the usual feed-forward
nets with sum-and-sigmoid units and one hidden layer. They were trained using the crossentropy error measure, with a desired output of one in a hand-chosen polygonal region
about each building cluster. The standard quadratic weight-decay term was added to this
to get the objective function, and only one regularization constant was used. This was
adjusted to give lowest error on a test set. We often found that the weights to two or more
hidden units would become identical and/or very small as the regularization term was
increased, and in these cases we pruned the extraneous units and began the search for the
optimal regularization parameter again. We usually ended up with very small networks,
with from two to five hidden units. Mter training a net at low-resolution, we expanded the
image of that net's output in order to use it as a context input for the networks at higher
resolution.
2.1.3 Performance
To compare the performance of the three systems, we chose a threshold for the networks'
outputs, and considered a building cluster to be detected by a network if the network's output was above the threshold at any pixel within the cluster. The number of detected clusters divided by the total number is the true-positive rate. The false-positive rate is more
difficult to define, since we had in mind an application in which the detection system
would draw a user's attention to objects which are likely to be of interest. The procedure
we used attempts to measure the amount of effort a human would expend in searching
regions with false detections by the network. See (Spence, et al., 1994) for details.
The performance figures presented here were measured on a validation set, i.e., an image
on which the network was not trained and which was not used to set the regularization
parameter. The results presented in Tables 1 and 2 are for a single threshold, chosen for
each network so that the true-positive rate was 90%. Figure 2 compares the ROC curves
of the three systems, i.e. the parametric curves of the true and false-positive rates as the
threshold on the network's output is varied. From Table 1, the features we used would
seem adequate for search at very low resolution, although the performance could be better.
Clay D. Spence, lo/m C. Pearson, Jim Bergen
986
Table 1: False-Positive Rates vs. Resolution.
These are results for the system with both road-map and context inputs.
PYRAMID LEVEL
FALSE-POSITIVE RATE
5
16%
3
4.6%
1
3.6%
Table 2: False-Positive Rates of the Search Systems at 90% True-Positive Rate.
NETWORK SYSTEM
FALSE-POSITIVE RATE
No Context or Road-Map Inputs
7.6%
Context Inputs, No Road Maps
5.8%
Context Inputs and Road Maps
3.6%
Table 2 and Figure 2 clearly show the benefits of using the road map and context inputs,
although the statistics are somewhat poor at the higher true-positive rates because of the
small number of building clusters which are being missed
2.2
LEARNING CONTEXT
Two things were changed to demonstrate that the context could be learned. First, the
unoriented spatial energy features are not well suited for distinguishing between roads and
other objects with a size similar to a road's width, so we used oriented energies instead.
These were the oriented energies derived using steerable filters (Freeman and Adelson,
1991) at four orientations. To force orientation imariance, we sorted the four oriented
energies at each pixel for each frequency band. These sorted oriented energy images were
then reduced in size as appropriate for the resolution being searched by a network. In this
case, we extracted energies only from the first, third, and fifth pyramid levels.
The second change is the use of hidden unit outputs for context inputs, instead of the output unit's outputs. The output units estimate the probability of finding a building cluster.
Although this may reflect information about roads, it is very indirect information about the
roads. It is more likely to carry some information about the coarse-scale appearance of the
potential building clusters. The hidden unit outputs should contain a richer description of
the image at a coarser scale.
2.2.1 Performance
The networks were trained in the same way as the networks described in Section 2.1. We
trained three networks to search levels five, three, and one. For comparison purposes, we
also trained a single-network to search in level one, with the same input features that were
used for all of the networks of the hierarchical search system. These include, for example,
three versions of each of the oriented energy images from level one (the second highest
frequency band). Two of these versions were reduced in size to levels three and five, and
then re-expanded to level one, so that they are simply blurred versions of the original
energy images. This gives the network a direct source of information on the coarse-scale
Coarse-To-Fine Image Search Using Neural Networks
987
100
Non-hierarchical
ID
95
.~
Hierarchical
.-- ????????? --..
"iii
o
a.
90
CIl
~
.=
~
85
80~~~~~~__~~~
a
10
15
5
% False Positives
20
Figure 3. Performance of Hierarchical and Non-hierarchical Algorithms that
Learn Context.
Figure 4. Image and Context Inputs to Highest-Resolution Network.
appearance of the image. although it is an expensive representation in terms of computation and memory. and the network has more inputs so that it takes longer to train.
The ROC curves for the two detection algorithms are shown in Figure 3. Their performance is about the same, suggesting that the coarser-scale networks in the hierarchical
system are not performing computations that are useful for the finest-scale network, rather
they simply pass on information about the coarse-scale appearance. In this example. the
advantage of the hierarchical system is in the efficiency of both the training of the algorithm and its use after training. Figure 4 shows the level-one image and the outputs of several hidden units from the level-three network, expanded to the same size. These outputs
suggest that road information is being extracted at coarse scale and passed to the high-resolution network.
3
DISCUSSION
We have performed "proof-of-concept" simulations on a realistic search problem of two of
the key concepts of our Hybrid Pyramid/Neural Network (HPNN) approach. We have
shown that: 1. The HPNN can learn to utilize explicitly-provided context data to improve
object detection; 2. the networks can learn to use context without being explicitly taught to
do so. This second point implies that the person training the system does not need to know
of the existence of the context objects. The networks at the different scales are trained in
sequence. starting at low resolution. A benefit of this coarse-to-fine training is a simple
form of exemplar selection. since we can choose to train a network only in the regions of
988
Clay D. Spence. John C. Pearson. Jim Bergen
interest as detected by the already-trained networks at lower resolution. We could also
train all of the networks simultaneously, so that the lower-resolution networks learn to
extract the most helpful information from the low-resolution image. The resulting system
would probably perform better, but it would also be more expensive to train.
This approach should work quite well for many automatic target recognition problems,
since the targets are frequently quite small. For more extended objects like clusters of
buildings, our method is an efficient way of examining the objects on several length
scales, but the detection at the finest scale is based on a small part of the potential object's
appearance. Extended images of real objects typically have many features at each of several resolutions. We are currently working on techniques for discovering several characteristic features at each resolution by training several networks at each resolution, and
integrating their responses to compute an overall probability of detection.
Acknowledgments
We would like to thank Peter Burt, P. Anandan, and Paul Sajda for many helpful discussions. This was work was funded by the National Information Display Laboratory and
under ARPA contract No. NOOOI4-93-C-0202.
References
D.H. Ballard and L.E. Wixson (1993) Object recognition using steerable filters at multiple
scales, Proceedings of the IEEE Workshop on Qualitative Vision, New York, NY.
D. Burr and Y. Miyata (1993) Hierarchical recurrent networks for learning musical structure, Proceedings of the IEEE Conference on Neural Networks for Signal Processing Ill,
C. Kamm, G. Kuhn, B. Yoon, S.Y. Kung, and R. Chellappa, eds., Piscataway, NJ,
pp. 216-225.
P. J. Burt and E. H. Adelson. (1983) The Laplacian pyramid as a compact image code.
IEEE Transactions, Vol. COM-31:4, April, pp. 532-540.
P.J. Burt (1988a) Smart sensing with a pyramid vision machine, Proceedings of the IEEE
Vol. 76, pp. 1006-1015.
PJ. Burt (1988b) Attention mechanisms for vision in a dynamic world, Proceedings of the
9th International Conference on Pattern Recognition, pp. 977-987.
PJ. Burt, J.R. Bergen, R. Kolczynski, R. Hingorani, W.A. Lee, A. Leung, J. Lubin, and H.
Shvayster (1989) Object tracking with a moving camera, Proceedings of the IEEE Workshop on Motion, Irvine.
W.T. Freeman and E.H. Adelson (1991) The design and use of steerable filters, IEEE
Trans. PAMI, 12:9, pp. 891-906.
S.H. Lane, J.C. Pearson, and R. Sverdlove (1992) Neural networks for classifying image
textures, Proceedings of the Government Applications of Neural Networks Conference,
Dayton, Ohio.
M.e. Mozer (1994) Neural network music composition by prediction: Exploring the benefits of psychoacoustic constraints and multiscale processing, to appear in Connection Science.
PART IX
APPLICATIONS
| 982 |@word version:8 simulation:1 decomposition:1 carry:1 contains:1 tuned:1 com:4 surprising:1 must:1 finest:2 john:4 realistic:1 visible:2 blur:3 v:1 half:1 fewer:1 selected:1 discovering:1 coarse:14 location:3 sigmoidal:1 five:3 height:1 constructed:5 direct:1 become:1 lowresolution:1 qualitative:1 combine:2 burr:3 manner:1 frequently:1 multi:2 freeman:2 automatically:1 kamm:1 actual:1 window:2 becomes:1 begin:1 discover:1 provided:7 didn:1 lowest:3 interpreted:1 developed:1 finding:4 nj:2 ended:1 certainty:2 rols:2 control:1 unit:12 appear:1 positive:13 before:1 limit:1 id:1 pami:1 might:1 chose:1 studied:1 limited:1 sarnoff:3 acknowledgment:1 thirty:1 camera:1 spence:6 implement:1 procedure:3 steerable:3 dayton:1 area:3 road:26 integrating:1 suggest:1 get:2 selection:3 context:45 map:17 center:1 primitive:1 rural:1 starting:3 attention:2 resolution:56 simplicity:1 searching:4 target:3 user:2 distinguishing:4 recognition:4 expensive:3 cspence:1 coarser:2 yoon:1 sverdlove:1 region:8 highest:2 mentioned:1 mozer:2 dynamic:1 trained:13 smart:1 efficiency:4 blurring:1 indirect:1 various:1 represented:1 train:6 sajda:1 chellappa:1 detected:3 approached:1 pearson:6 quite:2 richer:1 otherwise:1 ability:2 statistic:1 sequence:3 advantage:1 net:9 subtracting:1 relevant:2 combining:1 multiresolution:1 description:1 exploiting:1 cluster:12 object:22 help:2 illustrate:1 recurrent:1 measured:1 exemplar:3 nearest:1 received:5 implies:1 differ:1 kuhn:1 filter:3 human:1 government:1 adjusted:1 exploring:1 proximity:1 around:1 considered:1 roi:2 purpose:2 currently:2 infonnation:1 rough:1 clearly:1 always:1 gaussian:4 rather:1 greenspan:1 crossentropy:1 derived:2 greatly:2 detect:2 helpful:2 bergen:6 leung:1 integrated:1 typically:3 entire:1 hidden:7 expand:1 pixel:9 canceled:1 overall:1 orientation:2 ill:1 extraneous:1 spatial:2 aware:1 construct:1 sampling:2 identical:1 adelson:6 few:3 surroundings:1 oriented:5 preserve:1 simultaneously:1 national:1 attempt:1 detection:10 interest:4 weakness:1 capable:1 necessary:1 perfonnance:1 desired:1 re:1 maca:1 arpa:1 increased:1 ordinary:2 usefulness:1 examining:1 too:1 person:1 international:1 contract:1 lee:1 quickly:1 again:1 reflect:1 successively:1 choose:2 account:1 potential:3 suggesting:1 blurred:3 explicitly:7 performed:1 capability:1 accuracy:1 musical:1 characteristic:2 worth:1 finer:1 ed:1 definition:1 energy:10 frequency:5 pp:5 associated:2 proof:1 irvine:1 noooi4:1 clay:5 sophisticated:1 feed:2 higher:8 response:1 improved:2 april:1 box:1 until:1 hand:2 receives:1 working:1 multiscale:1 building:14 effect:1 contain:2 true:5 concept:2 regularization:4 laboratory:2 width:2 octave:1 demonstrate:6 motion:1 image:54 novel:1 ohio:1 began:1 sigmoid:1 rotation:1 unoriented:1 composition:2 ai:2 automatic:1 had:1 funded:1 moving:1 longer:1 etc:1 binary:4 additional:1 anandan:1 somewhat:2 redundant:1 signal:2 multiple:1 hpnn:3 faster:1 compensate:1 long:1 divided:1 y:1 laplacian:4 prediction:1 vision:3 represent:1 pyramid:24 achieved:1 whereas:1 fine:9 spacing:1 source:1 extra:1 probably:1 tend:1 thing:1 seem:2 call:1 near:1 leverage:1 iii:2 gave:1 architecture:1 identified:1 idea:4 passed:2 effort:1 lubin:1 peter:1 york:1 adequate:3 useful:3 amount:1 band:5 processed:1 reduced:3 per:1 vol:2 taught:1 key:1 four:2 demonstrating:2 enormous:1 threshold:4 pj:2 utilize:1 sum:1 almost:1 missed:1 draw:1 layer:1 distinguish:1 display:2 quadratic:1 refine:1 strength:1 scanned:1 constraint:1 lane:2 span:1 pruned:1 performing:1 expanded:3 relatively:2 piscataway:1 poor:1 aerial:4 smaller:1 invariant:1 taken:1 computationally:1 mechanism:1 know:2 mind:1 available:2 operation:1 hierarchical:11 appropriate:1 hingorani:1 existence:2 original:7 subsampling:1 include:1 music:4 exploit:2 giving:1 objective:1 added:1 already:1 strategy:1 parametric:1 dependence:1 usual:2 distance:2 thank:1 extent:1 length:1 code:1 providing:1 difficult:2 negative:1 design:1 perform:1 extended:3 jim:5 digitized:2 varied:1 burt:10 inferred:1 connection:1 learned:3 narrow:1 registered:1 trans:1 usually:2 pattern:3 eighth:1 memory:1 natural:1 hybrid:2 force:1 improve:3 extract:2 validation:1 sufficient:1 classifying:1 lo:1 course:1 changed:1 neighbor:1 taking:1 correspondingly:1 expend:1 absolute:1 fifth:4 benefit:8 curve:4 dimension:1 world:1 unaware:1 doesn:1 forward:2 made:3 author:1 far:1 transaction:1 compact:1 mter:1 don:1 terrain:1 search:29 spectrum:1 table:5 learn:10 ballard:2 expanding:2 miyata:3 improving:1 investigated:1 complex:1 constructing:2 did:1 psychoacoustic:1 linearly:1 paul:1 repeated:3 complementary:1 fig:1 roc:3 cil:1 ny:1 precision:1 sub:2 explicit:5 third:4 ix:1 sensing:1 decay:1 workshop:2 polygonal:1 false:8 texture:1 suited:1 photograph:4 simply:3 appearance:5 likely:4 tracking:1 extracted:4 sorted:2 man:1 change:1 typical:1 reducing:2 total:1 pas:1 invariance:1 meaningful:1 searched:4 arises:1 kung:1 relevance:1 perfonning:1 princeton:1 |
7,224 | 983 | On the Computational Utility of
Consciousness
Donald W. Mathis and Michael C. Mozer
mathis@cs.colorado.edu, mozer@cs.colorado.edu
Department of Computer Science and Institute of Cognitive Science
University of Colorado, Boulder
Boulder, CO 80309-0430
Abstract
We propose a computational framework for understanding and
modeling human consciousness. This framework integrates many
existing theoretical perspectives, yet is sufficiently concrete to allow
simulation experiments. We do not attempt to explain qualia (subjective experience), but instead ask what differences exist within
the cognitive information processing system when a person is conscious of mentally-represented information versus when that information is unconscious. The central idea we explore is that the contents of consciousness correspond to temporally persistent states
in a network of computational modules. Three simulations are described illustrating that the behavior of persistent states in the
models corresponds roughly to the behavior of conscious states
people experience when performing similar tasks. Our simulations
show that periodic settling to persistent (i.e., conscious) states improves performance by cleaning up inaccuracies and noise, forcing
decisions, and helping keep the system on track toward a solution.
1
INTRODUCTION
We propose a computational framework for understanding and modeling consciousness. Though our ultimate goal is to explain psychological and brain imaging data
with our theory, and to make testable predictions, here we simply present the framework in the context of previous experimental and theoretical work, and argue that
12
Donald Mathis, Michael C. Mozer
it is sensible from a computational perspective. We do not attempt to explain
qualia-subjective experience and feelings. It is not clear that qualia are amenable
to scientific investigation. Rather, our aim is to understand the mechanisms underlying awareness, and their role in cognition. We address three key questions:
? What are the preconditions for a mental representation to reach consciousness?
? What are the computational consequences of a representation reaching consciousness? Does a conscious state affect processing differently than an unconscious state?
? What is the computational utility of consciousness? That is, what is the computational role of the mechanism(s) underlying consciousness?
2
THEORETICAL FRAMEWORK
Modular Cognitive Architecture. We propose that the human cognitive architecture consists of a set of functionally specialized computational modules (e.g.,
Fodor, 1983). We imagine the modules to be organized at a somewhat coarse level
and to implement processes such as visual object recognition, visual word-form
recognition, auditory word and sound recognition, computation of spatial relationships, activation of semantic representations of words, sentences, and visual objects,
construction of motor plans, etc. Cognitive behaviors require the coordination of
many modules. For example, functional brain imaging studies indicate that there
are several brain areas used for different subtasks during cognitive tasks such as
word recognition (Posner & Carr, 1992).
Modules Have Mapping And Cleanup Processes. We propose that modules perform an associative memory function in their domain, and operate via a
two-stage process: a fast, essentially feedforward input-output mapping 1 followed
by a slower relaxation search (Figure 1). The computational justification for this
two stage process is as follows. We assume that, in general, the output space of a
module can represent a large number of states relative to the number of states that
are meaningful or well formed-i.e., states that are interpretable by other modules
or (for output modules) that correspond to sensible motor primitives. If we know
which representations are well-formed, we can tolerate an inaccurate feedforward
mapping, and "clean up" noise in the output by constraining it to be one of the
well-formed states. This is the purpose of the relaxation step: to clean up the
output of the feedforward step, resulting in a well-formed state. The cleanup process knows nothing about which output state is the best response to the input;
it acts solely to enforce well-formedness. Similar architectures have been used recently to model various neuropsychological data (Hinton & Shallice, 1991; Mozer &
Behrmann, 1990; Plaut & Shallice, 1993). The empirical motivation for identifying
consciousness with the results of relaxation search comes from studies indicating
that the contents of consciousness tend to be coherent, or well-formed (e.g., Baars,
1988; Crick, 1994; Damasio, 1989).
Persistent States Enter Consciousness. In our model, module outputs enter
consciousness if they persist for a sufficiently long time. What counts as long enough
lWe do not propose that this process is feedforward at the neural level. Rather, we
mean that any iterative refinement of the output over time is unimportant and irrelevant.
On the Computational Utility of Consciousness
13
relaxation search
feedforward mapping
.00 ??
Figure 1: Modules consist of two components.
is not yet determined, but in order to model specific psychological data, we will be
required to make this issue precise. At that time a specific commitment will need
to be made, and this commitment must be maintained when modeling further data.
An important property of our model is that there is no hierarchy of modules with
respect to awareness, in contrast to several existing theories that propose that access
to some particular module (or neural processing area) is required for consciousness
(e.g., Baars, 1988). Rather, information in any module reaches awareness simply by
persisting long enough. The persistence hypothesis is consistent with the theoretical
perspectives of Smolensky (1988), Rumelhart et al (1986), Damasio (1989), Crick
and Koch (1990), and others.
2.1
WHEN ARE MENTAL STATES CONSCIOUS?
In our framework, the output of any module will enter consciousness if it persists in
time. The persistence of an output state of a module is assured if: (1) it is a point
attractor of the relaxation search (i.e., a well-formed state), and (2) the inputs to
the module are relatively constant , i.e., they continue to be mapped into the same
attractor basin.
While our framework appears to make strong claims about the necessary and sufficient conditions for consciousness, without an exact specification of the modules
forming the cognitive architecture, it is lacking as a rigorous, testable theory. A
complete theory will require not only a specification of the modules, but will also
have to avoid arbitrariness in claiming that certain cognitive operations or brain
regions are modules while other are not. Ultimately, one must identify the neurophysiological and neuroanatomical properties of the brain that determine the
module boundaries (see Crick, 1994, for a promising step in this regard).
3
COMPUTATIONAL UTILITY OF CONSCIOUSNESS
For the moment, suppose that our framework provides a sensible account of awareness phenomena (demonstrating this is the goal of ongoing work.) If one accepts
this, and hence the notion that a cleanup process and the resulting persistent states
are required for awareness, questions about the role of cleanup in the model become quite interesting because they are equivalent to questions about the role of
the mechanism underlying awareness in cognition. One question one might ask is
whether there is computational utility to achieving conscious states. That is, does a
system that achieves persistent states perform better than a system that does not?
14
Donald Mathis, Michael C. Mozer
Does a system that encourages settling to well-formed states perform better than a
system that does not? We now show that the answer to this question is yes.
3.1
ADDITION SIMULATION
To examine the utility of cleanup, we trained a module to perform a simple multistep
cognitive task: adding a pair of two-digit numbers in three steps. 2 We tested the
system with and without cleanup and compared the generalization performance.
The network architecture (Figure 2) consists of a single module. The inputs consist
of the problem statement and the current partial solution-state. The output is
an updated solution-state. The module's output feeds back into its input. The
problem statement is represented by four pools of units, one for each digit of each
operand, where each pool uses a local encoding of digits. Partial solution states are
represented by five pools, one for each of the three result digits and one for each of
the two carry digits.
..........
(
~ ~ jU ~ if)
Projection
(copy)
1result 11
Iresult 31
Iresult21
tI
leany21
+
I
Input-ot,rtput
mapping
leany
(copy)
I
hidden units
"-
.,/
~[~E::~:::][:iIl[iI] Iresult111resutt 211result311eany 111eany 21
"
Figure 2: Network architecture for the addition task
Each addition problem was decomposed into three steps (Figure 3), each describing
a transformation from one partial solution state to the next, and the mapping net
was trained perform each transformation individually.
? ?
48
+ 62
???
step 1
--..
? 1
48
+ 62
??O
step 2
--..
+
1 1
48
62
?10
step 3
--..
+
1 1
48
62
11 0
Figure 3: The sequence of steps in an example addition problem
Step 1 Given the problem statement, activate the rightmost result digit and rightmost carry digit (comprising the first partial solution).
2 Of course, we don't believe that there is a brain module dedicated to addition problems.
This choice was made because addition is an intuitive example of a multistep task.
On the Computational Utility of Consciousness
15
Step 2 Given the first partial solution, activate the next result and carry digits
(second partial solution).
Step 3 Given the second partial solution, activate the leftmost result digit (final
solution).
The set of well-formed states in this domain consists of all possible combinations of
digits and "don't knows" among the pools ("don't knows" are denoted by question
marks in Figure 3) . Local representations of digits are used within each pool, and
"don't knows" are represented by the state in which no unit is active. Thus, the
set of well-formed states are those in which either one or no units are active in
each pool. To make these states attractors of the cleanup net, the connections
were hand-wired such that each pool was a winner-take-all pool with an additional
attractor at the zero state.
To run the net, a problem statement pattern is clamped on the input units, and
the net is allowed to update for 200 iterations. Unit activities were updated using
an incremental rule approximating continuous dynamics:
ai(t)
= TI(L: Wijaj(t -
1)) + (1 - T)ai(t - 1)
j
where ai(t) is the activity of unit i at time t, T is a time constant in the interval
[0,1]' and 10 is the usual sigmoid squashing function.
Figure 4 shows the average generalization performance of networks run with and
without cleanup, as a function of training set size. Note that, in principle, it is
not necessary for the system to have a cleanup process to learn the training set
perfectly, or to generalize perfectly. Thus, it is not simply the case that no solutions
exist without cleanup. The generalization results were that for any size training set,
percent correct on the generalization set is always better with cleanup than without.
This indicates that although the mapping network often generalizes incorrectly, the
output pattern often falls within the correct attractor basin. This is especially
beneficial in multistep tasks because cleanup can correct the inaccuracies introduced
by the mapping network, preventing the system from gradually diverging from the
desired trajectory.
-
=
Projection
No projection
training set size
(% of all problems)
Figure 4: Cleanup improves generalization performance.
Figure 5 shows an example run of a trained network. There is one curve for each
of the five result and carry pools, showing the degree of "activity" of the ultimate
target pattern, t, for that pool as a function of time. Activity is defined to be
e-lit-aIl2 where a is the current activity pattern and t is the target. The network
/6
Donald Mathis, Michael C. Mozer
solves the problem by passing though the correct sequence of intermediate states,
each of which are temporarily persistent. This resembles the sequence of conscious
states a person might experience while performing this task; each step of the problem
is performed by an unconscious process, and the results of each of step appear in
conscious awareness.
1.0r----=====~-----_""::--_~""-:-"""""'----------,
Activation of
target pattern
in each pool
.8
.6
:
.'
:'"
.'
,:'
result digit I,
- - carry digit I
"
"
result digit 2,
----- carrydigit2
...":'
,: ,'.
.......... result digit 3
- - - - - - - - - - ---".:!/
~~~~IO~~~~2~O~--~3~O---~4~O---~50~--~
TIme
Figure 5: Network solving the addition task in three steps
3.2
CHOICE POINT SIMULATION
In many ordinary situations, people are required to make decisions, e.g., drive
straight through an intersection or turn left, order macaroni or a sandwich for
lunch. At these choice points, any of the alternative actions are reasonable a priori.
Contextual information determines which action is correct, e.g., whether you are
trying to drive to work or to the supermarket. Conscious decision making often
occurs at these choice points, except when the task is overlearned (Mandler, 1975).
We modeled a simple form of a choice point situation. We trained a module to
output sequences of states, e.g., ABCD or EFGH, where states were represented by
unique activity patterns over a set of units. If the sequences shared no elements,
then presenting the first element of any sequence would be sufficient to regenerate
the sequence. But when sequences overlap, choice points are created. For example,
with the sequences ABCD and AEFG, state A can be followed by either B or E.
We show that cleanup allows the module to make a decision and complete one of
the two sequences. Figure 6 shows the operation of the module with and without
cleanup following presentation of an A after training on the sequence pair ABCD and
AEFG. There is one curve for each state, showing the activation of that state (defined
as before), as a function of time. When the network is run with cleanup, although
both states Band E are initially partially activated, the cleanup process maps this
ill-formed state to state B, and the network then correctly completes the sequence
ABCD. Without cleanup, the initial activation of states Band E causes a blending of
the two sequences ABCD and AEFG and the state degenerates. 3
Although the arithmetic and choice point tasks seem simple in part because we
predefined the set of well-formed states. However, because the architecture segre3In this simulation, we are not modeling the role of context in helping to select one
sequence or another; we are simply assuming that either sequence is valid in the current
context. The nature of the model does not change when we consider context. Assuming
that the domain is not highly overlearned, the context will not strongly evoke one alternative action or the other in the feedforward mapping, leading to partial activation of
multiple states, and the cleanup process will be needed to force a decision.
On the Computational Utility of Consciollsness
Activity
of state
0 .?
17
stateB
__ state C
- - - -state D
........ state E
o.
Figure 6: Decision-point task with and without cleanup
gates knowledge of well-formedness from knowledge of how to solve problems in the
domain, well-formedness could be learned simultaneously with, or prior to learning the task. One could imagine training the cleanup network to autoassociate
states it observes in the domain before or during training using an unsupervised or
self-supervised procedure.
4
COMPUTATIONAL CONSEQUENCES OF
PERSISTENT STATES
In a network of modules, a persistent well-formed state in one module exerts larger
influences on the state of other modules than do transient or ill-formed states. As a
result the dynamics of the system tends to be dominated by well-formed persistent
states. We show this in a final simulation.
The network consisted of two modules, A and B, connected in a simple feedforward
cascade. Each module's cleanup net was trained to have ten attractors, locally
represented in a winner-take-all pool of units. The mapping network of module B
was trained to map the attractors of module A to attractors in B in a one-to-one
fashion. Thus, state a1 in module A is mapped to {31 in module A, a2 to {32, etc.
Module B was initialized to a well-formed state {31, and the output state of module
A was varied, creating three conditions. In the persistent well-formed condition,
module A was clamped in the well-formed state a2 for 50 time steps. In the transient
well-formed condition, module A was clamped in state a2 for only 30 time steps. And
in the ill-formed condition, module A was clamped in an ill-formed state in which
two states, a2 and a3, were both partially active. Figure 7 shows the subsequent
activation of state {32 in module B as a function of time. Module B undergoes a
transition from state {31 to state {32 only in the persistent well-formed condition.
This indicates that the conjunction of well-formedness and persistence is required
to effect a transition from one state to another.
5
CONCLUSIONS
Our computational framework and simulation results suggest the following answers
to our three key questions:
18
Donald Mathis, Michael C. Mozer
well-formed, persistent
state <X2 in module A
Activity of
state ~2 in
module B
well-formed, transient
state <X2 in module A
0'
ill-formed, persistent
state in module A
Figure 7: Well-formeclness and persistence are both required for attractor transitions.
? In order to reach consciousness, the output of a module must be both persistent and semantically well-formed, and must not initiate an overlearned
process.
? The computational consequences of conscious (persistent) representations
include exerting larger influences on the cognitive system, resulting in increased ability to drive response processes such as verbal report.
? The computational utility of consciousness in our model lies in the ability of
cleanup to "focus" cognition, by keeping the system close to states which
are semantically meaningful. Because the system has learned to process
such states, performance is improved.
References
Baars, B. J. (1988) A Cognitive Theory of Con8ciou&ne8ll, Cambridge University Press.
Crick, F. (1994) The astonishing hypothesis: The scientific search for the soul. Scribner.
Crick, F., & Koch, C. (1990) Towards a neurobiological theory of consciousness. Sem.
Neuro., 2: 263-275
Damasio, A. (1989) The brain binds entities and events by multiregional activation from
convergence zones. Neural Computation, 1, 123-132
Fodor, J. A. (1983) The modularity of mind: An e8llay on faculty p8ychology. Cambridge,
MA: MIT Press.
Hinton, G. E., & Shallice, T. (1991) Lesioning an attractor network: Investigations of
acquired dyslexia., P8ych. Rev., 98: 74-95
Mandler, G. (1975) Consciousness: Respectable, useful and probably necessary. In Information Processing and Cognition, The Loyola Symposium, R. Solso (Ed.). Erlbaum.
Mozer, M. C., & Behrmann, M. (1990). On the interaction of spatial attention and lexical
knowledge: A connectionist account of neglect dyslexia. Cognitive Neuro8Cience, 2, 96-123.
Plaut, D. C., & Shallice, T. (1993) Perseverative and semantic influences on visual object
naming errors in optic aphasia: A connectionist account. J. Cog. Neuro.,5(1): 89-117
Posner, M. 1., & Carr, T. (1992) Lexical access and the brain: Anatomical constraints on
cognitive models of word recognition. American Journal of P8ychology, 105(1): 1-26
Rumelhart, D. E., Smolensky, P., McClelland, J. L., & Hinton, G. E. (1986) Schemata
and sequential thought in PDP models. In J. L. McClelland & D. E. Rumelhart (Eds.),
Parallel Di8tributed Proceuing, Vol. 2. Cambridge, MA: MIT Press.
Smolensky, P. (1988) On the proper treatment of connectionism. Brain Behav. Sci., 11:
1-74
| 983 |@word illustrating:1 faculty:1 simulation:8 carry:5 moment:1 initial:1 rightmost:2 subjective:2 existing:2 current:3 contextual:1 activation:7 yet:2 must:4 subsequent:1 motor:2 interpretable:1 update:1 mental:2 coarse:1 plaut:2 provides:1 five:2 become:1 symposium:1 persistent:16 consists:3 acquired:1 roughly:1 behavior:3 examine:1 brain:9 decomposed:1 underlying:3 what:6 transformation:2 scribner:1 act:1 ti:2 unit:9 appear:1 before:2 persists:1 local:2 bind:1 tends:1 consequence:3 io:1 encoding:1 proceuing:1 solely:1 multistep:3 might:2 resembles:1 co:1 neuropsychological:1 unique:1 implement:1 digit:15 procedure:1 area:2 empirical:1 cascade:1 thought:1 projection:3 persistence:4 word:5 donald:5 suggest:1 lesioning:1 close:1 context:5 influence:3 equivalent:1 map:2 lexical:2 primitive:1 attention:1 identifying:1 rule:1 posner:2 notion:1 justification:1 fodor:2 updated:2 target:3 hierarchy:1 unconscious:3 colorado:3 imagine:2 cleaning:1 construction:1 exact:1 suppose:1 us:1 hypothesis:2 element:2 rumelhart:3 recognition:5 persist:1 role:5 module:49 precondition:1 region:1 connected:1 observes:1 mozer:8 dynamic:2 ultimately:1 trained:6 solving:1 astonishing:1 differently:1 represented:6 various:1 fast:1 activate:3 quite:1 modular:1 larger:2 solve:1 ability:2 final:2 associative:1 sequence:15 net:5 propose:6 interaction:1 commitment:2 degenerate:1 intuitive:1 convergence:1 wired:1 incremental:1 object:3 solves:1 strong:1 c:2 indicate:1 come:1 correct:5 human:2 transient:3 require:2 generalization:5 investigation:2 formedness:4 connectionism:1 blending:1 helping:2 sufficiently:2 koch:2 iil:1 cognition:4 mapping:10 claim:1 achieves:1 a2:4 purpose:1 integrates:1 coordination:1 individually:1 mit:2 always:1 aim:1 rather:3 reaching:1 avoid:1 conjunction:1 focus:1 indicates:2 baars:3 contrast:1 rigorous:1 inaccurate:1 initially:1 hidden:1 comprising:1 issue:1 among:1 ill:5 denoted:1 priori:1 plan:1 spatial:2 lit:1 unsupervised:1 others:1 report:1 connectionist:2 di8tributed:1 simultaneously:1 attractor:10 sandwich:1 attempt:2 highly:1 activated:1 amenable:1 predefined:1 partial:8 necessary:3 experience:4 initialized:1 desired:1 theoretical:4 psychological:2 lwe:1 increased:1 modeling:4 respectable:1 ordinary:1 erlbaum:1 answer:2 periodic:1 person:2 ju:1 pool:12 michael:5 concrete:1 central:1 cognitive:13 creating:1 american:1 leading:1 account:3 performed:1 damasio:3 schema:1 parallel:1 formed:25 correspond:2 identify:1 yes:1 generalize:1 trajectory:1 drive:3 straight:1 explain:3 reach:3 ed:2 auditory:1 treatment:1 ask:2 knowledge:3 improves:2 exerting:1 organized:1 back:1 appears:1 feed:1 tolerate:1 supervised:1 response:2 improved:1 though:2 strongly:1 stage:2 hand:1 undergoes:1 scientific:2 believe:1 effect:1 consisted:1 hence:1 consciousness:22 semantic:2 during:2 self:1 encourages:1 maintained:1 leftmost:1 trying:1 presenting:1 complete:2 carr:2 dedicated:1 percent:1 recently:1 sigmoid:1 specialized:1 mentally:1 functional:1 operand:1 arbitrariness:1 overlearned:3 winner:2 functionally:1 cambridge:3 enter:3 ai:3 persisting:1 access:2 specification:2 etc:2 perspective:3 irrelevant:1 forcing:1 certain:1 continue:1 additional:1 somewhat:1 determine:1 arithmetic:1 ii:1 mandler:2 sound:1 multiple:1 long:3 naming:1 a1:1 prediction:1 neuro:2 essentially:1 exerts:1 iteration:1 represent:1 addition:7 interval:1 completes:1 ot:1 operate:1 probably:1 tend:1 seem:1 feedforward:7 constraining:1 enough:2 intermediate:1 affect:1 architecture:7 perfectly:2 idea:1 whether:2 utility:9 ultimate:2 passing:1 cause:1 behav:1 action:3 useful:1 clear:1 unimportant:1 conscious:10 band:2 ten:1 locally:1 mcclelland:2 exist:2 track:1 correctly:1 anatomical:1 vol:1 key:2 four:1 demonstrating:1 achieving:1 clean:2 imaging:2 relaxation:5 run:4 you:1 reasonable:1 decision:6 followed:2 perseverative:1 activity:8 optic:1 constraint:1 x2:2 dominated:1 performing:2 relatively:1 department:1 combination:1 beneficial:1 lunch:1 making:1 rev:1 gradually:1 boulder:2 describing:1 count:1 mechanism:3 turn:1 needed:1 know:5 initiate:1 mind:1 aphasia:1 generalizes:1 operation:2 enforce:1 dyslexia:2 alternative:2 slower:1 gate:1 neuroanatomical:1 include:1 neglect:1 testable:2 especially:1 approximating:1 question:7 occurs:1 usual:1 mapped:2 sci:1 entity:1 sensible:3 argue:1 toward:1 loyola:1 assuming:2 modeled:1 relationship:1 abcd:5 statement:4 claiming:1 shallice:4 proper:1 perform:5 regenerate:1 incorrectly:1 situation:2 hinton:3 precise:1 pdp:1 varied:1 subtasks:1 introduced:1 pair:2 required:6 sentence:1 connection:1 coherent:1 accepts:1 learned:2 inaccuracy:2 address:1 pattern:6 soul:1 smolensky:3 memory:1 overlap:1 event:1 settling:2 force:1 temporally:1 created:1 supermarket:1 prior:1 understanding:2 relative:1 lacking:1 interesting:1 versus:1 awareness:7 degree:1 basin:2 consistent:1 sufficient:2 principle:1 squashing:1 course:1 copy:2 keeping:1 verbal:1 allow:1 understand:1 institute:1 fall:1 regard:1 boundary:1 curve:2 valid:1 transition:3 preventing:1 made:2 refinement:1 feeling:1 neurobiological:1 keep:1 evoke:1 active:3 don:4 search:5 iterative:1 continuous:1 modularity:1 promising:1 learn:1 nature:1 sem:1 mathis:6 domain:5 assured:1 motivation:1 noise:2 nothing:1 allowed:1 solso:1 fashion:1 lie:1 clamped:4 behrmann:2 cog:1 specific:2 showing:2 a3:1 consist:2 adding:1 sequential:1 intersection:1 simply:4 explore:1 forming:1 neurophysiological:1 visual:4 temporarily:1 partially:2 corresponds:1 determines:1 ma:2 goal:2 presentation:1 towards:1 shared:1 crick:5 content:2 change:1 determined:1 except:1 semantically:2 experimental:1 diverging:1 meaningful:2 indicating:1 select:1 zone:1 people:2 mark:1 cleanup:23 ongoing:1 tested:1 phenomenon:1 |
7,225 | 984 | A Convolutional Neural Network
Hand Tracker
Steven J. Nowlan
Synaptics, Inc.
2698 Orchard Parkway
San Jose, CA 95134
nowlan@synaptics .com
John C. Platt
Synaptics, Inc.
2698 Orchard Parkway
San Jose, CA 95134
platt@synaptics.com
Abstract
We describe a system that can track a hand in a sequence of video
frames and recognize hand gestures in a user-independent manner.
The system locates the hand in each video frame and determines
if the hand is open or closed. The tracking system is able to track
the hand to within ?10 pixels of its correct location in 99.7% of
the frames from a test set containing video sequences from 18 different individuals captured in 18 different room environments. The
gesture recognition network correctly determines if the hand being
tracked is open or closed in 99.1 % of the frames in this test set .
The system has been designed to operate in real time with existing
hardware.
1
Introduction
We describe an image processing system that uses convolutional neural networks to
locate the position of a (moving) hand in a video frame, and to track the position of
this hand across a sequence of video frames. In addition , for each frame, the system
determines if the hand is currently open or closed. The input to the system is a
sequence of black and white, 320 by 240 pixel digitized video frames. We designed
the system to operate in a user-independent manner, using video frames from indoor
scenes with natural clutter and variable lighting conditions. For ease of hardware
implementation, we have restricted the system to use only convolutional networks
and simple image filtering operations, such as smoothing and frame differencing.
902
Steven J. Nowlan, John C. PIau
Figure 1: Average over all examples of each of the 10 classes of handwritten digits,
after first aligning all of the examples in each class before averaging.
Our motivation for investigating the hand tracking problem was to explore the limits
of recognition capability for convolutional networks. The structure of convolutional
networks makes them naturally good at dealing with translation invariance, and
with coarse representations at the upper layers, they are also capable of dealing
with some degree of size variation. Convolutional networks Qave been successfully
applied to machine print OCR (Platt et aI, 1992), machine print address block
location (Wolf and Platt, 1994), and hand printed OCR (Le Cun et aI, 1990; Martin
and Rashid, 1992). In each of these problems, convolutional networks perform very
well on simultaneously segmenting and recognizing two-dimensional objects.
In these problems, segmentation is often the most difficult step, and once accomplished the classification is simplified. This can be illustrated by examining the
average of all of the examples for each class after alignment and scaling. For the
case of hand-printed OCR (see Fig. 1), we can see that the average of all of the
examples is quite representative of each class, suggesting that the classes are quite
compact, once the issue of translation invariance has been dealt with. This compactness makes nearest neighbor and non-linear template matching classifiers reasonable
candidates for good performance.
If you perform the same trick of aligning and averaging the open and closed hands
from our training database of video sequences, you will see a quite different result
(Fig. 2). The extreme variability in hand orientations in both the open and closed
cases means that the class averages, even after alignment, are only weakly characteristic of the classes of open and closed hands. This lack of clean structure in
the class average images suggested that hand tracking is a challenging recognition
problem. This paper examines whether convolutional networks are extendable to
hand tracking, and hence possibly to other problems where classification remains
difficult even after segmentation and alignment.
2
System Architecture
The overall architecture of the system is shown in Fig. 3. There are separate hand
tracking and gesture recognition subsystems. For the hand tracking subsystem,
each video frame is first sub-sampled and then the previous video frame (stored) is
subtracted from the current video frame to produce a difference frame. These difference frames provide a crude velocity signal to the system, since the largest signals
in the difference frames tend to occur near objects that are moving (Fig. 5). Independent predictions of hand locations are made by separate convolutional networks,
which look at either the intensity frame or the difference frame. A voting scheme
then combines the predictions from the intensity and difference networks along with
predictions based on the hand trajectory computed from 3 previous frames.
A Convolutional Neural Network Hand Tracker
903
Figure 2: Average over all examples of open and closed hands from the database of
training video sequences, after first aligning all of the examples in each class before
averaging.
Previous
Subsampled
Video Frame
Current
Video
Frame
Difference
Hanel Locator
Network
Difference
Frame
Intensity
1-'----"'" Hand Locator
Position Hypotheses
Network
Position in
Previous Frames
Voting
Procedure
Hand PosHlan
~
Frame
Cropping
,.. Open/Close
Network
,
Is Hand Open?
Figure 3: Architecture of object recognition system for hand tracking and openversus-closed hand identification.
| 984 |@word hence:1 print:2 open:9 correct:1 illustrated:1 white:1 separate:2 existing:1 current:2 com:2 tracker:2 nowlan:3 image:3 differencing:1 john:2 difficult:2 tracked:1 designed:2 implementation:1 perform:2 currently:1 upper:1 ai:2 largest:1 successfully:1 rashid:1 coarse:1 variability:1 location:3 frame:24 locate:1 digitized:1 moving:2 synaptics:4 along:1 intensity:3 aligning:3 combine:1 manner:2 accomplished:1 address:1 able:1 suggested:1 captured:1 indoor:1 compactness:1 signal:2 video:14 pixel:2 classification:2 issue:1 orientation:1 overall:1 natural:1 gesture:3 smoothing:1 once:2 scheme:1 locates:1 prediction:3 look:1 voting:2 classifier:1 platt:4 addition:1 recognize:1 before:2 individual:1 simultaneously:1 segmenting:1 subsampled:1 limit:1 filtering:1 operate:2 tend:1 black:1 degree:1 alignment:3 challenging:1 extreme:1 ease:1 near:1 translation:2 architecture:3 capable:1 block:1 digit:1 neighbor:1 template:1 procedure:1 whether:1 printed:2 matching:1 made:1 san:2 simplified:1 close:1 subsystem:2 compact:1 recognizing:1 clutter:1 dealing:2 examining:1 investigating:1 hardware:2 parkway:2 stored:1 extendable:1 examines:1 track:3 correctly:1 ca:2 variation:1 user:2 containing:1 us:1 possibly:1 hypothesis:1 trick:1 clean:1 velocity:1 recognition:5 motivation:1 suggesting:1 database:2 steven:2 jose:2 fig:4 representative:1 you:2 inc:2 reasonable:1 sub:1 position:4 closed:8 scaling:1 candidate:1 environment:1 crude:1 layer:1 capability:1 weakly:1 occur:1 convolutional:10 characteristic:1 scene:1 dealt:1 handwritten:1 identification:1 trajectory:1 lighting:1 martin:1 describe:2 orchard:2 across:1 explore:1 quite:3 cun:1 tracking:7 naturally:1 restricted:1 sampled:1 wolf:1 determines:3 sequence:6 remains:1 segmentation:2 room:1 operation:1 averaging:3 ocr:3 invariance:2 subtracted:1 hand:29 cropping:1 produce:1 lack:1 object:3 locator:2 nearest:1 |
7,226 | 985 | Efficient Methods for Dealing with
Missing Data in Supervised Learning
Volker '!'resp?
Siemens AG
Central Research
Otto-Hahn-Ring 6
81730 Miinchen
Germany
Ralph Neuneier
Siemens AG
Central Research
Otto-Hahn-Ring 6
81730 Miinchen
Germany
Subutai Ahmad
Interval Research Corporation
1801-C Page Mill R<;l.
Palo Alto, CA 94304
Abstract
We present efficient algorithms for dealing with the problem of missing inputs (incomplete feature vectors) during training and recall.
Our approach is based on the approximation of the input data distribution using Parzen windows. For recall, we obtain closed form
solutions for arbitrary feedforward networks. For training, we show
how the backpropagation step for an incomplete pattern can be
approximated by a weighted averaged backpropagation step. The
complexity of the solutions for training and recall is independent
of the number of missing features. We verify our theoretical results
using one classification and one regression problem.
1
Introduction
The problem of missing data (incomplete feature vectors) is of great practical and
theoretical interest. In many applications it is important to know how to react if the
available information is incomplete, if sensors fail or if sources of information become
A.t the time of the research for this paper, a visiting researcher at the Center for
Biological and Computational Learning, MIT. E-mail: Volker.Tresp@zfe.siemens.de
690
VoLker Tresp. RaLph Neuneier. Subutai Ahmad
unavailable. As an example, when a sensor fails in a production process, it might
not be necessary to stop everything if sufficient information is implicitly contained
in the remaining sensor data. Furthermore, in economic forecasting, one might want
to continue to use a predictor even when an input variable becomes meaningless (for
example, due to political changes in a country). As we have elaborated in earlier
papers, heuristics such as the substitution of the mean for an unknown feature can
lead to solutions that are far from optimal (Ahmad and Tresp, 1993, Tresp , Ahmad,
and Neuneier, 1994). Biological systems must deal continuously with the problem
of unknown uncertain features and they are certainly extremely good at it. From
a biological point of view it is therefore interesting which solutions to this problem
can be derived from theory and if these solutions are in any way related to the way
that biology deals with this problem (compare Brunelli and Poggio, 1991). Finally,
having efficient methods for dealing with missing features allows a novel pruning
strategy: if the quality of the prediction is not affected if an input is pruned, we
can remove it and use our solutions for prediction with missing inputs or retrain
the model without that input (Tresp, Hollatz and Ahmad, 1995).
In Ahmad and Tresp (1993) and in Tresp, Ahmad and Neuneier (1994) equations for
training and recall were derived using a probabilistic setting (compare also Buntine
and Weigend, 1991, Ghahramani and Jordan, 1994). For general feedforward neural
networks the solution was in the form of an integral which has to be approximated using numerical integration techniques. The computational complexity of these
solutions grows exponentially with the number of missing features. In these two
publications, we could only obtain efficient algorithms for networks of normalized
Gaussian basis functions. It is of great practical interest to find efficient ways of
dealing with missing inputs for general feedforward neural networks which are more
commonly used in applications. In this paper we describe an efficient approximation
for the problem of missing information that is applicable to a large class of learning
algorithms, including feedforward networks. The main results are Equation 2 (recall) and Equation 3 (training). One major advantage of the proposed solution is
that the complexity does not increase with an increasing number of missing inputs.
The solutions can easily be generalized to the problem of uncertain (noisy) inputs.
2
2.1
Missing Information During Recall
Theory
We assume that a neural network N N(x) has been trained to predict E(ylx), the
expectation of y E !R given x E ~. During recall we would like to know the
(XC, XU) where XC
network's prediction based on an incomplete input vector x
denotes the known inputs and XU the unknown inputs. The optimal prediction
given the known features can be written as (Ahmad and Tresp, 1993)
=
Efficient Methods for Dealing with Missing Data in Supervised Learning
691
u
~=X
x;~,a)
Yi o
Xl:
.
?
00 :
o
000 o
o
X
c
Figure 1: The circles indicate 10 Gaussians approximating the input density distribution. XC = Xl indicates the known input, X2 = XU is unknown.
Similarly, for a network trained to estimate class probabilities, N Ni(X) ~
P(classilx), simply substitute P(classdx C) for E(ylx C ) and N Ni(X C , XU) for
N N (XC, XU) in the last equation.
The integrals in the last equations Can be problematic. In the worst case they have
to be approximated numerically (Tresp, Ahmad and Neuneier, 1994) which is costly,
since the computation is exponential in the number of missing inputs. For networks
of normalized Gaussians, there exist closed form solutions to the integrals (Ahmad
and Tresp, 1993). The following section shows how to efficiently approximate the
integral for a large class of algorithms.
2.2
An Efficient Approximation
Parzen windows are commonly used to approximate densities. Given N training
data {(xk, yk)lk = 1, ... , N}, we can approximate
1 N
P(x) ~ N L:G(x;xk,O')
(1)
k=l
where
k
11k 2
G( X; X ,0') = (211'0'2)D /2 exp( - 20'211x - X II )
is a multidimensional properly normalized Gaussian centered at data xk with variance 0'2. It has been shown (Duda and Hart (1973)) that Parzen windows approximate densities for N - 00 arbitrarily well, if 0' is appropriately scaled.
692
Volker Tresp, Ralph Neuneier, Subutai Ahmad
Using Parzen windows we may write
where we have used the fact that
and where G( xc; xc,le, u) is a Gaussian projected onto the known input dimensions
(by simply leaving out the unknown dimensions in the exponent and in the normalization, see Ahmad and Tresp, 1993). xc,le are the components of the training data
corresponding to the known input (compare Figure 1).
Now, if we assume that the network prediction is approximately constant over the
"width" of the Gaussians, u, we can approximate
J
NN(xC, x") G(XC,x";xle,u) dx"
~ NN(xC,x",Ie) G(XC;xc,le,u)
where N N(x C, x",Ie) is the network prediction which we obtain if we substituted the
corresponding components of the training data for the unknown inputs.
With this approximation,
(2)
Interestingly, we have obtained a network of normalized Gaussians which are
centered at the known components of the data points. The" output weights"
N N(xC, x",Ie) consist of the neural network predictions where for the unknown input
the corresponding components of the training data points have been substituted.
Note, that we have obtained an approximation which has the same structure as the
solution for normalized Gaussian basis functions (Ahmad and Tresp, 1994).
In many applications it might be easy to select a reasonable value for u using prior
knowledge but there are also two simple ways to obtain a good estimate for u using
leave-one-out methods. The first method consists of removing the k - th pattern
from the training data and calculating P(x le ) ~ N:l L:f:l,l# G(x le ; xl, u). Then
select the u for which the log likelihood L:1e log P(x le ) is maximum. The second
method consists of treating an input of the k - th training pattern as missing and
then testing how well our algorithm (Equation 2) can predict the target. Select the
u which gives the best performance. In this way it would even be possible to select
input-dimension-specific widths Ui leading to "elliptical", axis-parallel Gaussians
(Ahmad and Tresp, 1993).
Efficient Methods for Dealing with Missing Data in Supervised Learning
693
Note that the complexity of the solution is independent of the number of missing
inputs! In contrast, the complexity of the solution for feedforward networks suggested in Tresp, Ahmad and Neuneier (1994) grows exponentially with the number of
missing inputs. Although similar in character to the solution for normalized RBFs,
here we have no restrictions on the network architecture which allows us to choose
the network most appropriate for the application.
If the amount of training data is large, one can use the following approximations:
? Select only the K nearest data points. The distance is determined based
on the known inputs. K can probably be reasonably small ? 10). In the
extreme case, K = 1 and we obtain a nearest-neighbor solution. Efficient
tree-based algorithms exist for computing the K-nearest neighbors.
? Use Gaussian mixtures instead of Parzen windows to estimate the input
data distribution . Use the centers and variances of the components in
Equation 2.
? Use a clustering algorithm and use the cluster centers instead of the data
points in Equation 2.
Note that the solution which substitutes the components of the training data closest
to the input seems biologically plausible.
2.3
Experimental Results
We tested our algorithm using the same data as in Ahmad and Tresp, 1993. The
task was to recognize a hand gesture based on its 2D projection. As input, the
classifier is given the 2D polar coordinates of the five finger tip positions relative to
the 2D center of mass of the hand (the input space is therefore 10-D). A multi-layer
perceptron was trained on 4368 examples (624 poses for each gesture) and tested on
a similar independent test set. The inputs were normalized to a variance of one and
u was set to 0.1. (For a complete description of the task see (Ahmad and Tresp,
1993).) As in (Ahmad & Tresp, 1993) we defined a correct classification as one
in which the correct class was either classified as the most probable or the second
most probable. Figure 2 shows experimental results. On the horizontal axis, the
number of randomly chosen missing inputs is shown. The continuous line shows
the performance using Equation 2 where we used only the 10 nearest neighbors in
the approximation. Even with 5 missing inputs we obtain a score of over 90 %
which is slightly better than the solution we obtained in Ahmad and Tresp (1993)
for normalized RBFs. We expect our new solution to perform very well in general
since we can always choose the best network for prediction and are not restricted
in the architecture. As a benchmark we also included the case where the mean of
the missing input was substituted. With 5 missing inputs, the performance is less
than 60 %.
694
Volker Tresp. Ralph Neuneier. Subutai Ahmad
3D - Hanel Geetu", R8CC91ftIOn : 10 inpuIa
----.----
0.9
'.
0.8
'.
!II.
!.7
....
?..
.
" .
0.8
"
.
.?... ".
"
.
.?.. ......
0.5
2
3
4
5
8
runber 01 rristing Inputs
....,
7
Figure 2: Experimental results using a generalization data set. The continuous line
indicates the performance using our proposed method. The dotted lines indicate
the performance if the mean of the missing input variable is substituted. As a
comparison, we included the results obtained in Ahmad and Tresp (1993) using the
closed-form solution for RBF -networks (dashed).
3
Training (Backpropagation)
For a complete pattern (x.l:, yk), the weight update of a backpropagation step for
weight Wj is
k
A
?
(k _ NN.w ( x k))8NN
U.w
8 w (x ) .
J (X Y
Wj
Using the approximation of Equation 1, we obtain for an incomplete data point
(compare Tresp, Ahmad and Neuneier, 1994)
(3)
Here, IE compl indicates the sum over complete patterns in the training set, and (111
is the standard deviation of the output noise. Note that the gradient is a network
of normalized Gaussian basis functions where the "output-weight" is now
Efficient Methods for Dealing with Missing Data in Supervised Learning
695
The derivation of the last equation can be found in the Appendix. Figure 3 shows
experimental results.
Boston housing data: 13 inputs
0.15r------,r-------,r------,--~--__r--__._--"""T""--....,
0.14
0.13
sub6titute mean
0.12
1
B0.11
28 CQI11)Iete pattems
.....
--------------------~~--------------
0.1
0.09
0.08
O?061'----2'------J3'------J4'-----L5--~6-----'-7----'-8----'9
number of missing inputs
Figure 3: In the experiment, we used the Boston housing data set, which consists of
506 samples. The task is to predict the housing price from 13 variables which were
thought to influence the housing price in a neighborhood. The network (multilayer perceptron) was trained with 28 complete patterns plus an additional 225
incomplete samples. The horizontal axis indicates how many inputs were missing
in these 225 samples. The vertical axis shows the generalization performance. The
continuous line indicates the performance of our approach and the dash-dotted line
indicates the performance, if the mean is substituted for a missing variable. The
dashed line indicates the performance of a network only trained with the 28 complete
patterns.
4
Conclusions
We have obtained efficient and robust solutions for the problem of recall and training
with missing data. Experimental results verified our method. All of our results can
easily be generalized to the case of noisy inputs.
Acknowledgement
Valuable discussions with Hans-Georg Zimmermann, Tomaso Poggio, Michael Jordan and Zoubin Ghahramani are greatfully acknowledged. The first author would
like to thank the Center for Biological and Computational Learning (MIT) for providing and excellent research environment during the summer of 1994.
696
5
Volker Tresp, Ralph Neuneier, Subutai Ahmad
Appendix
Assuming the standard signal-plus-Gaussian-noise model we obtain for a complete
sample
P(xk,ykl{wi})
=G(yk;NNw(xk),(1y) P(x k)
where {Wi} is the set of weights in the network. For an incomplete sample
Using the same approximation as in Section 2.2,
p(xc,k, ykl{ W;}) ~
L
G(yk; N Nw(xc,k, xu,I), (1y) G(xc,k; xc, I, (1)
l?compl
where I sums over all complete samples. As before, we substitute for the missing components the ones from the complete training data. The log-likelihood C
(a function of the network weights {Wi}) can be calculated as (x k can be either
complete or incomplete) C
Lf=llog p(xk, yk I{ Wi}). The maximum likelihood
solution consists of finding weights {Wi} which maximize the log-likelihood. Using
the approximation of Equation 1, we obtain for an incomplete sample as gradient
Equation 3 (compare Tresp, Ahmad and Neuneier, 1994).
=
References
Ahmad, S. and Tresp, V. (1993). Some Solutions to the Missing Feature Problem in Vision.
In S. J. Hanson, J. D. Cowan and C. L. Giles, (Eds.), Advances in Neural Information
Processing Systems 5, San Mateo, CA: Morgan Kaufmann .
Brunelli, R. and Poggio, T. (1991). HyperBF Networks for Real Object Recognition. IJCAL
Buntine, W. L. and Weigend, A. S. (1991). Bayesian Back-Propagation. Complex systems,
Vol. 5, pp. 605-643.
Duda, R. O. and Hart, P. E. (1973). Pattern Classification and Scene Analysis.
Wiley and Sons, New York.
John
Ghahramani, Z. and Jordan, M. 1. (1994) . Supervised Learning from Incomplete Data via
an EM approach. In: Cowan, J. D., Tesauro, G., and Alspector, J., eds., Advances in
Neural Information Processing Systems 6, San Mateo, CA, Morgan Kaufman.
Tresp, V., Ahmad, S. and Neuneier, R. (1994). Training Neural Networks with Deficient Data. In: Cowan, J. D., Tesauro, G., and Alspector, J., eds., Advances in Neural
Information Processing Systems 6, San Mateo, CA, Morgan Kaufman.
Tresp, V., Hollatz, J. and Ahmad, S. (1995). Representing Probabilistic Rules with Networks of Gaussian Basis Functions. Accepted for publication in Machine Learning.
| 985 |@word duda:2 seems:1 substitution:1 score:1 interestingly:1 neuneier:12 elliptical:1 dx:1 must:1 written:1 john:1 numerical:1 remove:1 treating:1 update:1 xk:6 miinchen:2 five:1 become:1 consists:4 alspector:2 tomaso:1 multi:1 window:5 increasing:1 becomes:1 alto:1 mass:1 kaufman:2 ag:2 finding:1 corporation:1 multidimensional:1 scaled:1 classifier:1 before:1 approximately:1 might:3 plus:2 mateo:3 averaged:1 practical:2 testing:1 lf:1 backpropagation:4 thought:1 projection:1 zoubin:1 onto:1 influence:1 restriction:1 missing:29 center:5 zfe:1 react:1 rule:1 coordinate:1 resp:1 target:1 approximated:3 recognition:1 worst:1 wj:2 ahmad:27 valuable:1 yk:5 environment:1 complexity:5 ui:1 trained:5 basis:4 easily:2 finger:1 derivation:1 describe:1 neighborhood:1 heuristic:1 plausible:1 otto:2 noisy:2 housing:4 advantage:1 description:1 cluster:1 ring:2 leave:1 object:1 pose:1 nearest:4 b0:1 indicate:2 correct:2 centered:2 everything:1 generalization:2 biological:4 probable:2 exp:1 great:2 nw:1 predict:3 major:1 polar:1 applicable:1 palo:1 weighted:1 mit:2 subutai:5 sensor:3 gaussian:8 always:1 volker:6 publication:2 derived:2 properly:1 indicates:7 likelihood:4 political:1 contrast:1 nn:4 nnw:1 pattems:1 greatfully:1 germany:2 ralph:5 classification:3 exponent:1 integration:1 having:1 biology:1 randomly:1 recognize:1 interest:2 certainly:1 mixture:1 extreme:1 integral:4 necessary:1 poggio:3 tree:1 incomplete:11 circle:1 theoretical:2 uncertain:2 earlier:1 giles:1 deviation:1 predictor:1 buntine:2 density:3 ie:4 l5:1 probabilistic:2 tip:1 parzen:5 continuously:1 michael:1 central:2 choose:2 ykl:2 leading:1 de:1 view:1 hollatz:2 closed:3 parallel:1 rbfs:2 elaborated:1 ni:2 variance:3 kaufmann:1 efficiently:1 bayesian:1 researcher:1 classified:1 ed:3 pp:1 stop:1 recall:8 knowledge:1 back:1 supervised:5 furthermore:1 hand:2 horizontal:2 propagation:1 quality:1 grows:2 verify:1 normalized:9 brunelli:2 deal:2 during:4 width:2 generalized:2 complete:9 novel:1 exponentially:2 numerically:1 similarly:1 j4:1 han:1 closest:1 tesauro:2 continue:1 arbitrarily:1 yi:1 morgan:3 additional:1 maximize:1 dashed:2 ii:2 signal:1 gesture:2 hart:2 prediction:8 j3:1 regression:1 multilayer:1 vision:1 expectation:1 normalization:1 want:1 interval:1 source:1 country:1 leaving:1 appropriately:1 meaningless:1 probably:1 deficient:1 cowan:3 jordan:3 feedforward:5 easy:1 architecture:2 economic:1 forecasting:1 york:1 ylx:2 amount:1 exist:2 problematic:1 dotted:2 write:1 vol:1 affected:1 georg:1 acknowledged:1 verified:1 sum:2 weigend:2 reasonable:1 appendix:2 layer:1 dash:1 summer:1 x2:1 scene:1 extremely:1 pruned:1 slightly:1 son:1 character:1 em:1 wi:5 biologically:1 restricted:1 zimmermann:1 equation:13 fail:1 know:2 available:1 gaussians:5 appropriate:1 substitute:3 denotes:1 remaining:1 clustering:1 xc:17 calculating:1 ghahramani:3 hahn:2 approximating:1 strategy:1 costly:1 visiting:1 gradient:2 distance:1 thank:1 mail:1 assuming:1 providing:1 unknown:7 perform:1 vertical:1 benchmark:1 arbitrary:1 hyperbf:1 xle:1 hanson:1 suggested:1 pattern:8 including:1 representing:1 lk:1 axis:4 tresp:27 prior:1 acknowledgement:1 relative:1 expect:1 interesting:1 sufficient:1 production:1 last:3 perceptron:2 neighbor:3 dimension:3 calculated:1 author:1 commonly:2 projected:1 san:3 far:1 compl:2 pruning:1 approximate:5 implicitly:1 dealing:7 continuous:3 reasonably:1 robust:1 ca:4 unavailable:1 excellent:1 complex:1 substituted:5 main:1 noise:2 xu:6 retrain:1 wiley:1 fails:1 position:1 exponential:1 xl:3 removing:1 specific:1 consist:1 boston:2 mill:1 simply:2 contained:1 rbf:1 price:2 change:1 included:2 determined:1 llog:1 accepted:1 experimental:5 siemens:3 select:5 tested:2 |
7,227 | 986 | Grammar Learning by a Self-Organizing
Network
Michiro Negishi
Dept. of Cognitive and Neural Systems, Boston University
111 Cummington Street
Boston, MA 02215 email: negishi@cns.bu.edu
Abstract
This paper presents the design and simulation results of a selforganizing neural network which induces a grammar from example sentences. Input sentences are generated from a simple phrase
structure grammar including number agreement, verb transitivity, and recursive noun phrase construction rules. The network
induces a grammar explicitly in the form of symbol categorization
rules and phrase structure rules.
1 Purpose and related works
The purpose of this research is to show that a self-organizing network with a certain
structure can acquire syntactic knowledge from only positive (i.e. grammatical)
data, without requiring any initial knowledge or external teachers that correct
errors.
There has been research on supervised neural network models of language acquisition tasks [Elman, 1991, Miikkulainen and Dyer, 1988, John and McClelland, 1988].
Unlike these supervised models, the current model self-organizes word and phrasal
categories and phrase construction rules through mere exposure to input sentences,
without any artificially defined task goals. There also have been self-organizing
models of language acquisition tasks [Ritter and Kohonen, 1990, Scholtes, 1991].
Compared to these models, the current model acquires phrase structure rules in
more explicit forms, and it learns wider and more structured contexts, as will be
explained below.
2 Network Structure and Algorithm
The design of the current network is motivated by the observation that humans
have the ability to handle a frequently occurring sequence of symbols (chunk)
as an unit of information [Grossberg, 1978, Mannes, 1993]. The network consists
of two parts: classification networks and production networks (Figure 1). The
classification networks categorize words and phrases, and the production networks
28
Michiro Negishi
evaluate how it is likely for a pair of categories to form a phrase. A pair of combined
categories is given its own symbol, and fed back to the classifiers.
After weights are formed, the network parses a sentence as follows. Input words
are incrementally added to the neural sequence memory called the Gradient Field
[Grossberg, 1978] (GF hereafter). The top (i.e. most recent) two symbols and the
lookahead token are classified by three classification networks. Here a symbol
is either a word or a phrase, and the lookahead token is the word which will be
read in next. Then the lookahead token and the top symbol in the GF are sent to
the right production network, and the top and the second ones are sent to the left
production network. If the latter pair is judged to be more likely to form a phrase,
the symbol pair reduces to a phrase, and the phrase is fed back to the GF after
removing the top two symbols. Otherwise, the lookahead token is added to the
sequence memory, causing a shift in the sequence memory. If the input sentence is
grammatical, the repetition of this process reduces the whole sentence to a single
"5" (sentence) symbol. The sequence of shifts and reductions (annoted with the
resultant symbols) amounts to a parse of the sentence.
During learning, the operations stated above are carried out as weights are gradually formed. In classification networks, the weights record a distribution pattern
with respect to each symbol. That is, the weights record the co-occurrence of
up to three adjacent symbols in the corpus. An symbol is classified in terms of
this distribution in the classification networks. The production networks keep
track of the categories of adjacent symbols. If the occurrence of one category reliably predicts the next or the previous one, the pair of categories forms a phrase,
and is given the status of an symbol which is treated just like a word in the
sentence. Because the symbols include phrases, the learned context is wider
and more structured than the mere bigram, as well as the contexts utilized in
[Ritter and Kohonen, 1990, Scholtes, 1991].
3 Simulation
3.1 The Simulation Task
The grammar used to generate input sentences (Table 3) is identical to that used
in [Elman,1991], except that it does not include optionally transitive verbs and
proper nouns. Lengths of the input sentences are limited to 16 words. To determine the completion of learning, after accepting 200 consecutive sentences with
learning, learning is suppressed and other 200 sentences are processed to see if all
are accepted. In addition, the network was tested for 44 ungrammatical sentences
to see that they are correctly rejected. Ungrammatical sentences are derived by
hand from randomly generated grammatical sentences. Parameters used in the
simulation are : number of symbol nodes = 30 (words) + 250 (phrases), number
of category nodes = ISO, f. = 10- 9, 'Y = 0.25, p = 0.65, 0'1 = 0.00005, /31 = 0.005,
f32 = 0.2, 0'3 = 0.0001, /33 = 0.001, and T = 4.0.
Grammar Learning by a Self-Organizing NeMork
29
3.2 Acquired Syntax Rules
Learning was completed after learning 19800 grammatical sentences. Tables 1 and
2 show the acquired syntax rules extracted from the connection weights. Note that
category names such as Ns, VPp, are not given a priori, but assigned by the author
for the exposition. Only rules that eventually may reach the "S"(sentence) node
are shown. There were a small number of uninterpretable rules, which are marked
I/?". These rules might disturb normal parsing for some sentences, but they were
not activated while testing for 200 sentences after learning.
3.3 Discussion
Recursive noun phrase structures should be learned by finding equivalences of
distribution between noun phrases and nouns. However, nouns and noun phrases
have the same contextual features only when they are in certain contexts. An
examination of the acquired grammar reveals that the network finds equivalence
of features not of ''N'' and ''N RC" (where RC is a relative clause) but of "N V" and
''N RC V" (when ''N RC" is subjective), or "V N" and /IV NRC" (when ''N RC"
is objective). As an example, let us examine the parsing of the sentence [19912]
below. The rule used to reduce FEEDS CATS WHO UVE (''V NRC") is PO, which
is classified as category C4, which includes P121 (''V N") where V are the singular
forms of transitive verbs, and also includes the ''V'' where V are singular forms of
intransitive verbs. Thus, GIRL WHO FEEDS CATS WHO UVE is reduced to GIRL
WHO "VPsingle".
***[19912}***********??**********??*******??************ ?? *******.****
+---141--- +
I
+---88------+
I
I
+---206------+
I
I
I
+----0----+
I
I
I
I
+-219-+
I
I
+-41-+
I
+-36-+
I
BOYS CHASE GIRL WHO FEEDS CATS WHO LIVE
?Accepted?
Top symbol was 77
4 Conclusion and Future Direction
In this paper, a self-organizing neural network model of grammar learning was
presented. A basic principle of the network is that all words and phrases are
categorized by the contexts in which they appear, and that familiar sequence of
categories are chunked.
As it stands, the scope of the grammar used in the simulation is extremely limited.
Also, considering the poverty of the actual learning environment, the learning
of syntax should also be guided by the cognitive competence to comprehend the
utterance situations and conversational contexts. However, being a self-organizing
network, the current model offers a plausible model of natural language acquisition
through mere exposures to only grammatical sentences, not requiring any external
teacher or an explicit goal.
30
Miclziro Negislzi
Table 1. Acquired categorization rules
S
:=
C4
:a
C13
C16
C18
C20
C26
.:=
..-
:=
C29
.-
C30
:=
C32
.-
C29 r NPs VPs ., I
C30r??,1
enr~vpp?,
LIVES I ALKS I
POrvrsNpRC?' I
P74 rvrsNs RC?, I
P121 rvrsNs?' I
P157 rvrsNpo,
GIRL I DOG I
CAT I BOY
CHASE I FEED
WHO
CHASES I FEEDS
BOYS I CATS I
DOGS I GIRLS
P93 r Ns RC VPs?' I
P138 rNs VPs?'
P2 vrp NP\.iPp ., I
P94 r vrp N
., I
P137 r??,
WALK I LIVE I
PI r~NpRC?' I
P61
pNpo, I
P88rvrpNs RC?, I
P122 r vrp Ns?'
r
=rR?,
= rvrs?'
=rNp?'
P41 rNsR?,
.- P36 rNpR?,
C69
.-
C74
.-
C77
.-
C1l9
C122
C139
:=
.. rVps?'
= /"Ns?,
= rvrp?'
.-
C52
C56
C58
= rNPsvps?,
wnere
=r? 0,
RCs
RCp
NPp
NPs
:=
..-
=
=
=
=
P28 r Ns vrs ., I
P34 r Np vrVi I
P68 r Ns RC so, I
P147 rNpRCvrp?,
P206 rNs R VPs?'
P238 rNsRNvr?,
P219 rNpRVpp
I
P249 rNpRNvr?,
P141 rNpVPp?' I
P217 rNpRCVpp?,
Pl48
P243
PIO rvrs NPs VPs?' I
P32 r vrs NPp VPp ? ,
0'
= rNvr?,
rNsRCs?'
=
= rNpRCp?'
= rNPpVPp?,
=/"vrsNvr?,
= rNsR vrsNvr?,
= /" VPs' VPp's
RVPs I RNvr
RVPp I RNvr
Np I NpRCp
Ns I NsRCs
r
= rvpp?'
Table 2. Acquired production rules
PO
P1
P2
PI0
P28
P32
P34
P36
P41
P61
P68
P74
P88
P93
P94
P121
P122
P137
P138
P141
P147
P148
P157
P206
P217
P219
P238
P243
P249
:=C20 rvrs?'
:=C16 rvrp?'
:= C16 rvrp?,
:= C20 rvrs?'
:=C13 rNs?'
:=C20 rvrs?'
:=C26 rNp?'
:= C26 ,. Np .,
:=C13 rNs?'
:= C16 rvrp?'
:=C69 rNsRCs?,
:=C20 rvrs?'
:=C16 rVTp?'
:= C69 rNsRCs?,
:= C16 rvrp?'
:= C20 r vrs ? ,
:=C16 rVTp?'
:= C122 r NsR vrsNVT?,
:=C13 rNs?'
:= C26 rNp?'
:=C74 rNp RCs?,
:= C20 rvrs
:=C20 rvrs?'
:= C52 r Ns R .,
:= C74 rNp RCs?,
:= C56 r Np R .,
:=C52rNsR?,
:= C52 r Ns R
:=C56 rNpR?,
0'
0'
0'
C74 rNpRCp
C74rNpRCp?'
en /"NPpVpp?'
C29 rNPs VPs?'
C20 rvrso,
C77/"NPpVPp?'
C16/"vrp?'
C18 /"R?I
C18 /"R?,
C26 /"Np?1
C20 /"vrs?'
C69 rNs RCs?,
C69 r Ns RCs .,
C4 r VPs?'
C58/"Nvr?,
C13/"Ns?,
C13 rNs?'
C32 /"VPp
C4rVps?'
C32 /"Vpp?,
C16 rvrp?'
C58/"Nvr?1
C26/"Np?'
C4 rVPs?1
C32r Vpp?'
C32r Vpp?'
C58 rNvrol
C119/"vrsNvr?,
C58/"Nvr?,
0'
= r vrs Np RCp .,
= rvrpNpRCp?'
= rvrpNPp VPp?'
.. rvrsNPs VPs?'
= rNsvrs?'
= /"vrsNPpVPp?'
= /"Npvrp?'
=rNpR?,
=rNsR?,
=/"vrpNp?'
= /"NsRCs vrs?'
= vrs Ns RCs .,
= rvrpNsRCs?'
=r Ns RCs VPs .,
= rvrpNvr?,
= /"vrs Ns?1
= rvrpNs?,
= /"??I
= rNsvps,
=/"NpVPp?'
=rNp RCs vrp?'
= rvrsNvr?,
= rvrsNp?,
= /"NsRVPs?'
=r Np RCs VPp .,
= /"NpRVPp?'
=/"NsRNvr?,
= /" (Ns R vrs N) vr ?1
= /"Np RNvr?,
r
?,
Grammar Learning by a Self-Organizing Network
3/
Acknowledgements
The author wishes to thank Prof. Dan Bullock, Prof. Cathy Harris, Prof. Mike
Cohen, and Chris Myers of Boston University for valuable discussions.
This work was supported in part by the Air Force Office of Scientific Research
(AFOSR F49620-92-J-0225).
References
[Elman, 1991] Elman, J. (1991). Distributed representations, simple recurrent networks, and grammatical structure. Machine Learning, 7.
[Grossberg, 1978] Grossberg, S. (1978). A theory of human memory: Selforganization and performance of sensory-motor codes, maps, and plans. Progress
in Theoretical Biology,S.
(John and McClelland, 1988] John, M. F. S. and McClelland, J. L. (1988). Applying
contextual constraints in sentence comprehension. In Touretzky, D. 5., Hinton,
G. E., and Sejnowsky, T. J., editors, Proceedings of the Second Connectionist Models
Summer School 1988, Los Altos, CA. Morgan Kaufmann Publisher, Inc.
[Mannes, 1993] Mannes, C. (1993). Self-organizing grammar induction using a
neural network model. In Mitra, J., Cabestany, J., and Prieto, A., editors, New
Trends in Neural Computation: Lecture Notes in Computer Science 686. Springer
Verlag, New York.
[Miikkulainen and Dyer, 1988] Miikkulainen, R. and Dyer, M. G. (1988). Encoding
input/output representations in connectionist cognitive systems. In Touretzky,
D. D., Hinton, G. E., and Senowsky, T. J., editors, Proceedings of the Second Connectionist Models Summer School 1988, Los Altos, CA. Morgan Kauffman Publisher,
Inc.
[Ritter and Kohonen, 1990] Ritter, H. and Kohonen, T. (1990). Learning seman totopic maps from context. Proceedings of. ITCNN 90, Washington D.C., 1.
[Scholtes, 1991] Scholtes, J. C. (1991). Unsupervised context learning in natural
language processing. Proceedings of I[CNN Seattle 1991.
Appendix A. Activation and learning equations
A.l Classification Network Activities
eGradient Field
(1 )
where t is a discrete time, i is the symbol id. and Ii(t) is an input symbol.
eInput Layer
X1Ai(t)
= O(2(XOi(t)-O(XOi(t)))),
X1Bi(t)
= O(XOi(t)),
X1ci(t)
= Ii(t+1)
Where the suffix A ,B, and C the most recent, the next to most recent, and the
lookahead symbols, respectively'. Weights in networks A, B, and C are identical.
O(x) =
1
if x > 1.-2- M
1. 0
otherwIse
J
32
Michiro Negishi
Here M is the maximum number of symbols on the gradient field .
? Feature Layer
X2;i
=L
Xl"j Wl"ji, X2;{1
= I(Xl;d(a+ 2: X2;j)),
j
X2"i
= X2;{ /(a+ 2: X2;f)
j
j
I(x) = 2/(1 + exp( -Tx)) - 1
where s is a suffix which is either A, B, or C and T is the steepness of the sigmoid
function and a is a small positive constant. Table 4 shows the meaning of above
suffix i .
? Category Layer
X3pi
I
={
0
if i = min{jl2:k8X2"kW2"kj > p}, or
if ? = min{jl2:k" X2"k W2"kj > p} & unreli =ja:r: {unrefJ}
otherwise
(2)
Where p is the least match score required and ure Ii is an unreferenced count.
A.2 Classification Learning
.Feature Weights
LlWl"ij = -alWl"ij +,81Xl i (X2"j - Wl"ij)
where al is the forgetting rate, and ,81 is the learning rate .
? Categorization Weights
{ Ll W2"ij = /32X3"i(X2"i - W2"ij) if the node is selected by the first line of (2)
W2"ij = X2"i
if the node is selected by the second line of (2)
where /32 is the learning rate.
A.3 Production Network Activities
.Mutual predictiveness
=
=
X3Ai W3ij,
X3B j W4 ji ,
= X3cjW4 ji ,
= X3Bi W3 ij,
= X4ij XSji
= X7ij X8ji
The phrase identification number for a category pair (i, j) is given algOrithmically
in the current version by a cash function cash(i, j).
(i) Case in which 'Y 2:ij X6ij ~ 2:ij X9ij
X10 j
XOi(t
={
1
0
if i
= cash(I, J)
Reduce
where X6IJ =i]a:r: (X6ij )
otherwise
+ 1) = 0.5 * pop(pOp(XOi(t))) + XlO,
pop(x) = 2(x - 8(x))
Case in which 'Y Eij X6ij < 2:ij X9ij
Shift
The next input symbol is added on the gradient field, as was expressed in (1).
(ii)
Grammar Learning by a Self-Organizing Network
33
A.4 Production Learning
where X3 Ai and X3 Bj are nodes that receive the next to the most recent symbol i
and the most recent symbol j, respectively.
Shift I Reduce
Controller
Production
Networks
(predlctlveness
evaluators)
Classification ~ ~
Networks
... ~
(~
00
eedback
hrases
)
0
171o~_70v'ovO~I--+----.J
Neural Sequence Me ory
Lookahead
Token
Input words
Figure 1. Block diagram of the network
NP
VP
RC
N
N I NRC
V[NP]
whoNPV I whoVP
boy I girl I cat I dog I
boys I girls I cats 1 dogs
V
_
chase I feed I work I live I
chases I feeds I works I lives
urn
agreement
- Agreements between N and V within clause
- Agreements between head Nand
subordinate V (where a ro riate)
er arguments
- chase, feed -> require a direct object
- walk, live -> preclude a direct object
(Observed also for head/verb relations
in relative clauses)
Table 3. Grammar for generated sentences
Table 4. Subfields in a feature layer
34
Michiro Negishi
Category
(The MOst RlI08nt)
Category
(The "Next to
Most Recent)
To the
Production
r---~------------~~------------------------------"'Ne~k
X 3B
Category
(Lookahead)
o
C
Xo
(X,o>
'----A--""I.."'::. Gradient Field
Nonterm inal s
(Reduction
Results)
From the Production Network
Lookahead
Terminals
Terminals(lnputs)
... ABC ...
- - - - -.... Classification Path
c===J
terminals
- - -..
~ Copy 'or Learning
E:::;:;:@:?:~:@
nonterminals
category
_
prllYious
next
next to next
Figure 2. Classification Network
x
'0
Nonterminal
X 3C
Xa
V
Xg
+
X7
The ext to
Most Recent
Category
The Most Recent
Category
Figure 3. Production Network
Lookahead
Category
| 986 |@word cnn:1 version:1 selforganization:1 f32:1 bigram:1 simulation:5 reduction:2 initial:1 score:1 hereafter:1 subjective:1 current:5 contextual:2 activation:1 parsing:2 john:3 motor:1 selected:2 iso:1 record:2 accepting:1 node:6 evaluator:1 rc:10 scholtes:4 direct:2 consists:1 dan:1 acquired:5 forgetting:1 elman:4 frequently:1 examine:1 p1:1 terminal:3 pio:1 actual:1 preclude:1 considering:1 p28:2 alto:2 c13:6 finding:1 rcp:2 ro:1 classifier:1 unit:1 appear:1 positive:2 mitra:1 ext:1 encoding:1 id:1 ure:1 path:1 p219:2 might:1 equivalence:2 co:1 limited:2 kw2:1 subfields:1 grossberg:4 testing:1 recursive:2 block:1 x3:3 w4:1 word:10 judged:1 context:8 live:4 applying:1 map:2 exposure:2 rule:13 handle:1 phrasal:1 construction:2 agreement:4 trend:1 utilized:1 predicts:1 mike:1 observed:1 valuable:1 environment:1 girl:7 po:2 cat:7 tx:1 cathy:1 plausible:1 pi0:1 otherwise:4 grammar:13 ability:1 syntactic:1 chase:6 sequence:7 rr:1 myers:1 kohonen:4 causing:1 organizing:9 lookahead:9 los:2 seattle:1 disturb:1 categorization:3 object:2 wider:2 recurrent:1 completion:1 nonterminal:1 ij:10 school:2 progress:1 p2:2 xoi:5 direction:1 guided:1 correct:1 human:2 unreferenced:1 subordinate:1 ja:1 require:1 comprehension:1 predictiveness:1 normal:1 exp:1 negishi:5 scope:1 bj:1 p141:2 consecutive:1 purpose:2 wl:2 repetition:1 cash:3 office:1 derived:1 chunked:1 suffix:3 nand:1 p36:2 relation:1 classification:10 c16:9 priori:1 plan:1 noun:7 mutual:1 field:5 washington:1 identical:2 biology:1 unsupervised:1 future:1 np:15 connectionist:3 randomly:1 familiar:1 poverty:1 cns:1 intransitive:1 nprc:1 activated:1 iv:1 walk:2 theoretical:1 phrase:19 ory:1 nonterminals:1 teacher:2 combined:1 chunk:1 bu:1 ritter:4 cognitive:3 external:2 includes:2 inc:2 explicitly:1 formed:2 air:1 kaufmann:1 who:7 vp:1 identification:1 mere:3 vrs:9 classified:3 c32:3 reach:1 touretzky:2 email:1 acquisition:3 sejnowsky:1 resultant:1 knowledge:2 back:2 feed:8 nvr:3 supervised:2 just:1 rejected:1 xa:1 hand:1 parse:1 incrementally:1 scientific:1 name:1 nsr:1 requiring:2 assigned:1 npp:2 read:1 adjacent:2 comprehend:1 transitivity:1 self:10 during:1 ll:1 acquires:1 cummington:1 syntax:3 meaning:1 sigmoid:1 clause:3 ji:3 cohen:1 ai:1 alwl:1 language:4 own:1 recent:8 certain:2 verlag:1 life:2 morgan:2 determine:1 ii:4 reduces:2 x10:1 match:1 offer:1 rcs:9 basic:1 controller:1 receive:1 addition:1 p32:2 diagram:1 singular:2 publisher:2 w2:4 unlike:1 sent:2 w3:1 reduce:3 shift:4 motivated:1 michiro:4 york:1 selforganizing:1 amount:1 induces:2 processed:1 mcclelland:3 category:19 reduced:1 generate:1 algorithmically:1 track:1 correctly:1 discrete:1 steepness:1 appendix:1 uninterpretable:1 layer:4 summer:2 activity:2 constraint:1 x2:10 x7:1 argument:1 extremely:1 c20:10 conversational:1 min:2 urn:1 structured:2 rnp:6 suppressed:1 bullock:1 explained:1 gradually:1 xo:1 equation:1 eventually:1 count:1 dyer:3 fed:2 operation:1 jl2:2 occurrence:2 top:5 include:2 completed:1 prof:3 c74:4 objective:1 added:3 gradient:4 thank:1 prieto:1 street:1 chris:1 me:1 evaluate:1 induction:1 length:1 code:1 acquire:1 optionally:1 boy:5 stated:1 design:2 reliably:1 proper:1 observation:1 situation:1 hinton:2 head:2 rn:7 verb:5 competence:1 c26:6 pair:6 dog:4 required:1 sentence:24 connection:1 c4:4 learned:2 pop:3 below:2 pattern:1 kauffman:1 c18:3 including:1 memory:4 treated:1 examination:1 natural:2 force:1 ne:1 carried:1 xg:1 transitive:2 utterance:1 gf:3 kj:2 acknowledgement:1 relative:2 afosr:1 par:1 lecture:1 c30:1 ovo:1 principle:1 editor:3 pi:1 production:12 token:5 supported:1 copy:1 distributed:1 grammatical:6 f49620:1 stand:1 sensory:1 author:2 ipp:1 miikkulainen:3 enr:1 status:1 vps:10 keep:1 reveals:1 corpus:1 table:7 ca:2 p93:2 ungrammatical:2 artificially:1 whole:1 nrc:3 categorized:1 eedback:1 en:1 c52:3 vr:1 n:15 explicit:2 wish:1 xl:3 learns:1 removing:1 er:1 symbol:25 p41:2 occurring:1 vpp:10 boston:3 c29:3 eij:1 likely:2 expressed:1 xlo:1 springer:1 extracted:1 ma:1 harris:1 inal:1 abc:1 goal:2 marked:1 exposition:1 except:1 called:1 accepted:2 organizes:1 latter:1 categorize:1 dept:1 tested:1 |
7,228 | 987 | Recurrent Networks:
Second Order Properties and Pruning
Morten With Pedersen and Lars Kai Hansen
CONNECT, Electronics Institute
Technical University of Denmark B349
DK-2800 Lyngby, DENMARK
emails:with.lkhansen@ei.dtu.dk
Abstract
Second order properties of cost functions for recurrent networks
are investigated. We analyze a layered fully recurrent architecture,
the virtue of this architecture is that it features the conventional
feedforward architecture as a special case. A detailed description of
recursive computation of the full Hessian of the network cost function is provided. We discuss the possibility of invoking simplifying
approximations of the Hessian and show how weight decays iron the
cost function and thereby greatly assist training . We present tentative pruning results, using Hassibi et al.'s Optimal Brain Surgeon,
demonstrating that recurrent networks can construct an efficient
internal memory.
1
LEARNING IN RECURRENT NETWORKS
Time series processing is an important application area for neural networks and
numerous architectures have been suggested, see e.g. (Weigend and Gershenfeld, 94).
The most general structure is a fully recurrent network and it may be adapted using
Real Time Recurrent Learning (RTRL) suggested by (Williams and Zipser, 89). By
invoking a recurrent network, the length of the network memory can be adapted to
the given time series, while it is fixed for the conventional lag-space net (Weigend
et al., 90). In forecasting, however, feedforward architectures remain the most
popular structures; only few applications are reported based on the Williams&Zipser
approach. The main difficulties experienced using RTRL are slow convergence and
674
Morten With Pedersen, Lars Kai Hansen
lack of generalization. Analogous problems in feedforward nets are solved using
second order methods for training and pruning (LeCun et al., 90; Hassibi et al.,
92; Svarer et al., 93). Also, regularization by weight decay significantly improves
training and generalization. In this work we initiate the investigation of second order
properties for RTRL; a detailed calculation scheme for the cost function Hessian is
presented, the importance of weight decay is demonstrated, and preliminary pruning
results using Hassibi et al.'s Optimal Brain Surgeon (OBS) are presented. We find
that the recurrent network discards the available lag space and constructs its own
efficient internal memory.
1.1
REAL TIME RECURRENT LEARNING
The fully connected feedback nets studied by Williams&Zipser operate like a state
machine, computing the outputs from the internal units according to a state vector
z(t) containing previous external inputs and internal unit outputs. Let x(t) denote
a vector containing the external inputs to the net at time t, and let y(t) denote a
vector containing the outputs of the units in the net. We now arrange the indices
on x and y so that the elements of z(t) can be defined as
,
k EI
, kEU
where I denotes the set of indices for which Zk is an input, and U denotes the set of
indices for which Zk is the output of a unit in the net. Thresholds are implemented
using an input permanently clamped to unity. The k'th unit in the net is now
updated according to
where Wkj denotes the weight to unit k from input/unit j and
function of the k'th unit.
"'0 is the activation
When used for time series prediction, the input vector (excluding threshold) is
usually defined as x(t) = [x(t), . .. , x(t - L + 1)] where L denotes the dimension of
the lag space. One of the units in the net is designated to be the output unit Yo, and
its activating function 10 is often chosen to be linear in order to allow for arbitrary
dynamical range. The prediction of x(t + 1) is x(t + 1) = lo[so(t?). Also, if the first
prediction is at t = 1, the first example is presented at t = 0 and we 'set y(O) = O.
We analyse here a modification of the standard Williams&Zipser construction that
is appropriate for forecasting purposes. The studied architecture is layered. Firstly,
we remove the external inputs from the linear output unit in order to prevent the
network from getting trapped in a linear mode. The output then reads
x(t + 1)
=Yo(t + 1) = L
WojYj(t) + Wthres,o
(1)
jeU
=
=
Since y(O)
0 we obtain a first prediction yielding x(l)
Wthres,o which is likely
to be a poor prediction, and thereby introducing a significant error that is fed
back into the network and used in future predictions. Secondly, when pruning
Recurrent Networks: Second Order Properties and Pruning
675
a fully recurrent feedback net we would like the net to be able to reduce to a
simple two-layer feedforward net if necessary. Note that this is not possible with
the conventional Williams&Zipser update rule, since it doesn't include a layered
feedforward net as a special case. In a layered feedforward net the output unit is
disconnected from the external inputs; in this case, cf. (1) we see that x(t + 1) is
based on the internal 'hidden' unit outputs Yk(t) which are calculated on the basis
of z(t - 1) and thereby x(t -1). Hence, besides the startup problems, we also get
a two-step ahead predictor using the standard architecture.
In order to avoid the problems with the conventional Williams&Zipser update
scheme we use a layered updating scheme inspired by traditional feedforward nets,
in which we distinguish between hidden layer units and the output unit. At time t,
the hidden units work from the input vector zh(t)
,
,
k EI
kE U
, k=O
where I denotes the input indices, U denotes the hidden layer units and 0 the
output unit. Further, we use superscripts hand 0 to distinguish between hidden
unit and output units. The activation of the hidden units is calculated according
to
y~(t) = fr[s~(t)] = fr [
L
Wki
zJ (t)]
, kEU
(2)
ie1uUuO
The hidden unit outputs are forwarded to the output unit, which then sees the
input vector zkCt)
, kEU
OCt) _ { y~(t)
Zk
-
yO(t-1)
k=O
and is updated according to
(3)
The cost function is defined as C = E + wTRw. R is a regularization matrix, w is
the concatenated set of parameters, and the sum of squared errors is
1
E
T
= 2 L[e(t)F
, e(t)
= x(t) -
yO(t),
(4)
t=l
where T is the size of the training set series. RTRL is based on gradient descent in
the cost function, here we investigate accelerated training using Newton methods.
For that we need to compute first and second derivatives of the cost function. The
essential difficulty is to determine derivatives of the sum of squared errors:
aE = _ {-.. e(t) ayO(t)
aw?'3?
L...J
aw'3..
t=l
(5)
676
Morten With Pedersen, Lars Kai Hansen
The derivative of the output unit is computed as
8yO(t)
-8Wij
8r[sO(t)] 8s0(t)
._8so(t)
8Wij
(6)
where
8s0(t) _ 1: . O(t) "" . 8yjl(t)
-8-- - UO,Zj + L- WOJI 8
Wij
j/EU
Wij
+ woo 8
yO(t - 1)
8
Wij
(7)
where 6j k is the Kronecker delta. This expression contains the derivative of the
hidden units
(8)
where
(9)
...
...
132
132
<>
<>
~!:'::.3-.a~.2:;:-5---:'.a.'::-2---:.a7..15'--.a-:':.1'--.a7..0:;:-5-~-::0~.05~-::-0.1;---::'!0.1?5
WEIGHT VAlUE
~.3
.a.25
.a.2
.a.15
.a.1
.a.OS
WEIGHT VALUE
0.05
0.1
0.15
Figure 1: Cost function dependence of a weight connecting two hidden units for
the sunspot benchmark series. Left panel: Cost function with small weight decay,
the (local) optimum chosen is marked by an asterix. Right panel: The same slice
through the cost function but here retrained with higher weight decay.
The complexity of the training problem for the recurrent net using RTRL is demonstrated in figure 1. The important role of weight decay (we have used a simple weight
decay R = at) in controlling the complexity of the cost function is evident in the
right panel of figure 1. The example studied is the sunspot benchmark problem
(see e.g. (Weigend et al., 90) for a definition). First, we trained a network with
the small weight decay and recorded the left panel result. Secondly, the network
was retrained with increased weight decay and the particular weight connecting
two hidden units was varied to produce the right panel result. In both cases all
other weights remained fixed at their optimal values for the given weight decay. In
addition to the complexity visible in these one-parameter slices of the cost function, the cost function is highly anisotropic in weight space and consequently the
network Hessian is ill-conditioned. Hence, gradient descent is hampered by slow
con vergen ce.
Recurrent Networks; Second Order Properties and Pruning
2
677
SECOND ORDER PROPERTIES OF THE COST
FUNCTION
To improve training by use of Newton methods and for use in OBS-pruning we
compute the second derivative of the error functional:
t
8 2E
= _
[e(t) 8 2yO(t) _ 8 yO(t) . 8 yO (t)]
8Wij8wpq
t=l
8Wij8wpq
8Wij
8wpq
(10)
The second derivative of the output is
2
8wij 8wpq
8 yO(t) _
--....:.....:.-
8 2 r[sO(t)] 8s0(t) 8s0(t) 8r[sO(t)] 8 2 SO(t)
.- - .- - +
. ----:...;....;..8so(t)2
8Wij
8wpq
8so(t)
8Wij8wpq
(11)
with
8 2so(t) _, 8zJ(t) ~
8 2yj,(t)
Ooi+
~
Woj'
8 Wij 8 Wpq
OWpq
j'EU
8wij 8wpq
+ woo
8 2yO(t - 1)
8wij 8wpq
8z~(t)
+ Dop 8Wij
(12)
This expression contains the second derivative of the hidden unit outputs
8 2yi(t) _ 8 2 fr[si(t)] . 8si(t) . 8si(t) + 8fr[si(t)]. 02si(t)
OWijOWpq - osi(t)2
OWij
OWpq
osi(t)
OWijOWpq
(13)
with
02si(t) _
ozj(t) ~
02yj,(t - 1)
02 yO(t - 1)
oz~(t) (14)
- Dki -0-- + L..J Wkj
+
Wko
+
Dkp 0
8WijOWpq
Wpq j'EU
OWijOWpq
OWijOWpq
Wij
I
Recursion in the five index quantity (14) imposes a significant computational burden; in fact the first term of the Hessian in (10), involving the second derivative, is
often neglected for computational convenience (LeCun et al., 90). Here we start by
analyzing the significance of this term during training. We train a layered architecture to predict the sunspot benchmark problem. In figure 2 the ratio between the
largest eigenvalue of the second derivative term in (10) and the largest eigenvalue
of the full Hessian is shown. The ratio is presented for two different magnitudes of
weight decay. In line with our observations above the second order properties of the
"ironed" cost function are manageable, and we can simplify the Hessian calculation by neglecting the second derivative term in (10), i.e., apply the Gauss-Newton
approximation.
3
PRUNING BY THE OPTIMAL BRAIN SURGEON
Pruning of recurrent networks has been pursued by (Giles and Omlin, 94) using
a heuristic pruning technique, and significant improvement in generalization for a
sequence recognition problem was demonstrated. Two pruning schemes are based
on systematic estimation of weight saliency: the Optimal Brain Damage (OBD)
scheme of (LeCun et al., 90) and OBS by (Hassibi et al., 93). OBD is based
on the diagonal approximation of the Hessian and is very robust for forecasting
(Svarer et al., 93). If an estimate of the full Hessian is available OBS can be used
678
Morten With Pedersen, Lars Kai Hansen
10'
..:.::.: .... :::::;:;:::... :::: .
. '"
10...'----!-10--f::20--:30~----!
..:---~50--~---:::70:----=1O
..
ITERATION.
1040''-----:'10--f::20--:30~----!..:---~50,----..
~---:70~---:!IO
ITERATION.
Figure 2: Ratio between the largest magnitude eigenvalue of the second derivative
term of the Hessian (c.f. equation (10)) and the largest magnitude .eigenvalue of
the complete Hessian as they appeared during ten training sessions. The connected
circles represent the average ratio. Left panel: Training with small weight decay.
Right panel: Training with a high weight decay.
for estimation of saliencies incorporating linear retraining. In (Hansen and With
Pedersen, 94) OBS was generalized to incorporate weight decays; we use these
modifications in our experiments. Note that OBS in its standard form only allows
for one weight to be eliminated at a time. The result of a pruning session is a
nested family of networks. In order to select the optimal network within the family
it was suggested in (Svarer et al., 93.) to use the estimated test error. In particular
we use Akaike's Final Prediction Error (Akaike, 69) to estimate the network test
error Etest = ?(T + N)/(T - N? . 2E/T 1, and N is the number of parameters
in the network. In figure 3 we show the results of such a pruning session on the
sunspot data starting from a (4-4-1) network architecture. The recurrent network
was trained using a damped Gauss-Newton scheme. Note that the training error
increases as weights are eliminated, while the test error and the estimated test error
both pass through shallow minima showing that generalization is slightly improved
by pruning. In fact, by retraining the optimal architecture with reduced weight
decay both training and test errors are decreased in line with the observations in
(Svarer et al., 93). It is interesting to observe that the network, though starting
with access to a lag-space of four delay units, has lost three of the delayed inputs;
hence, rely solely on its internal memory, as seen in the right panel of. figure 3. To
further illustrate the memory properties of the optimal network, we show in figure
4 the network response to a unit impulse. It is interesting that the response of the
network extends for approximately 12 time steps corresponding to the "period" of
the sunspot series.
lThe use of Akaike's estimate is not well justified for a feedback net, test error estimates
for feedback models is a topic of current research.
Recurrent Networks: Second Order Properties and Pruning
679
OUTPUT
0.25 .
~0.'5
w
0.1 ,'
: .. "': -
.- - ... .~:.. __ ....
~ ..;
....
~
....... ._~_ . __
~ ...... ~~ .-: :7
0.05
~~-='0--~'5~~~~~~~~~--~$~-~~~"~~
X(I-I)
NUMSER OF PARAMETERS
X(I-2)
X(I-3)
X(I-4)
Figure 3: Left panel: OBS pruning of a (4-4-1) recurrent network trained on sunspot
benchmark. Development of training error, test error, and Akaike estimated test
error (FPE). Right panel: Architecture of the FPE-optimal network. Note that the
network discards the available lag space and solely predicts from internal memory.
0.8
0.1
! ..
0.5
0.4
.,
0.8
.-..
,
..
0.3
w
~
...
0.2
II!
0.'
.0.'
.0.20
I
. \ .,:
V"'0
15
TIME
~
,
\ :
orf"
t
.0.8
..\./
,
.0.1
...
:
"
I
.0.4
I
I
.'
:\
I
,,
,,
,--
'0
TIME
'5
~
Figure 4: Left panel: Output of the pruned network after a unit impulse input at
t = O. The internal memory is about 12 time units long which is, in fact, roughly
the period of the sunspot series. Right panel: Activity of the four hidden units in
the pruned network after a unit impUlse at time t = O.
4
CONCLUSION
A layered recurrent architecture, which has a feedforward net as a special case, has
been investigated. A scheme for recursive estimation of the Hessian of the fully
recurrent neural net is devised . It's been shown that weight decay plays a decisive
role when adapting recurrent networks. Further, it is shown that the' second order
information may be used to train and prune a recurrent network and in this process
the network may discard the available lag space. The network builds an efficient
680
Morten With Pedersen, Lars Kai Hansen
internal memory extending beyond the lag space that was originally available.
Acknowledgments
We thank J an Larsen, Sara Solla, and Claus Svarer for useful discuss~ons, and Lee
Giles for providing us with a preprint of (Giles and amlin, 94). We thank the
anonymous reviewers for valuable comments on the manuscript. This research is
supported by the Danish Natural Science and Technical Research Councils through
the Computational Neural Network Center (CONNECT).
References
H. Akaike: Fitting Autoregressive M ode/s for Prediction. Ann. Inst. Stat. Mat.
21, 243-247, (1969).
Y. Le Cun, J.S. Denker, and S.A. Solla: Optimal Brain Damage. In Advances
in Neural Information Processing Systems 2, (Ed. D.S. Touretzsky) Morgan Kaufmann, 598-605, (1990).
C.L. Giles and C.W. amlin: Pruning of Recurrent Neural Networks for Improved
Generalization Performance. IEEE Transactions on Neural Networks, to appear.
Preprint NEC Research Institute (1994).
L.K. Hansen and M. With Pedersen: Controlled Growth of Cascade Correlation
Nets, International Conference on Artificial Neural Networks ICANN'94 Sorrento.
(Eds. M. Marinaro and P.G. Morasso) Springer, 797-801, (1994).
B. Hassibi, D. G. Stork, and G. J. Wolff, Optimal Brain Surgeon and General
Network Pruning, in Proceedings of the 1993 IEEE International Conference on
Neural Networks, San Francisco (Eds. E.H. Ruspini et al. ) IEEE, 293-299 (1993).
C. Svarer, L.K. Hansen, and J. Larsen: On Design and Evaluation of .Tapped Delay
Line Networks, In Proceedings ofthe 1993 IEEE International Conference on Neural
Networks, San Francisco, (Eds . E.H. Ruspini et al. ) 46-51, (1993).
A.S . Weigend, B.A. Huberman, and D.E. Rumelhart: Predicting the future: A
Connectionist Approach. Int. J. of Neural Systems 3, 193-209 (1990).
A.S. Weigend and N.A. Gershenfeld, Eds.: Times Series Prediction: Forecasting the
Future and Understanding the Past. Redwood City, CA: Addison-Wesley (1994).
R.J. Williams and D. Zipser: A Learning Algorithm for Continually Running Fully
Recurrent Neural Networks, Neural Computation 1, 270-280, (1989).
| 987 |@word manageable:1 retraining:2 orf:1 simplifying:1 invoking:2 thereby:3 electronics:1 series:8 contains:2 past:1 current:1 activation:2 si:6 visible:1 remove:1 update:2 pursued:1 keu:3 firstly:1 five:1 fitting:1 roughly:1 brain:6 inspired:1 provided:1 wki:1 panel:12 ooi:1 growth:1 unit:34 uo:1 appear:1 continually:1 local:1 io:1 analyzing:1 fpe:2 solely:2 approximately:1 studied:3 sara:1 range:1 acknowledgment:1 lecun:3 yj:2 recursive:2 lost:1 area:1 significantly:1 adapting:1 cascade:1 get:1 convenience:1 layered:7 conventional:4 demonstrated:3 reviewer:1 center:1 williams:7 starting:2 ke:1 rule:1 analogous:1 updated:2 construction:1 controlling:1 play:1 akaike:5 tapped:1 element:1 rumelhart:1 recognition:1 updating:1 predicts:1 role:2 preprint:2 solved:1 connected:2 eu:3 solla:2 valuable:1 yk:1 complexity:3 neglected:1 trained:3 surgeon:4 basis:1 wtrw:1 train:2 artificial:1 startup:1 woj:1 lag:7 kai:5 heuristic:1 forwarded:1 analyse:1 superscript:1 final:1 sequence:1 eigenvalue:4 net:19 fr:4 osi:2 oz:1 description:1 yjl:1 getting:1 convergence:1 optimum:1 extending:1 produce:1 owij:1 illustrate:1 recurrent:24 stat:1 implemented:1 lars:5 activating:1 generalization:5 investigation:1 preliminary:1 anonymous:1 secondly:2 etest:1 predict:1 arrange:1 purpose:1 estimation:3 hansen:8 council:1 largest:4 city:1 avoid:1 yo:12 improvement:1 greatly:1 inst:1 hidden:12 wij:13 ill:1 development:1 special:3 construct:2 eliminated:2 future:3 connectionist:1 simplify:1 few:1 delayed:1 possibility:1 investigate:1 highly:1 evaluation:1 yielding:1 damped:1 neglecting:1 necessary:1 circle:1 increased:1 giles:4 cost:15 introducing:1 predictor:1 delay:2 reported:1 connect:2 aw:2 international:3 systematic:1 lee:1 asterix:1 connecting:2 squared:2 recorded:1 containing:3 external:4 derivative:11 int:1 decisive:1 analyze:1 start:1 kaufmann:1 saliency:2 ofthe:1 pedersen:7 ed:5 email:1 definition:1 danish:1 jeu:1 larsen:2 con:1 popular:1 improves:1 iron:1 back:1 manuscript:1 wesley:1 higher:1 originally:1 response:2 improved:2 though:1 correlation:1 hand:1 ei:3 a7:2 o:1 lack:1 mode:1 impulse:3 regularization:2 hence:3 read:1 during:2 generalized:1 evident:1 complete:1 functional:1 stork:1 anisotropic:1 significant:3 session:3 access:1 own:1 discard:3 touretzsky:1 yi:1 seen:1 minimum:1 morgan:1 prune:1 determine:1 period:2 ii:1 full:3 technical:2 calculation:2 long:1 devised:1 controlled:1 prediction:9 involving:1 ae:1 iteration:2 represent:1 justified:1 addition:1 ode:1 decreased:1 wkj:2 operate:1 claus:1 comment:1 zipser:7 feedforward:8 architecture:12 reduce:1 expression:2 assist:1 forecasting:4 hessian:12 useful:1 detailed:2 ten:1 reduced:1 zj:3 trapped:1 delta:1 estimated:3 mat:1 four:2 amlin:2 demonstrating:1 threshold:2 prevent:1 gershenfeld:2 ce:1 sum:2 weigend:5 extends:1 family:2 ob:7 layer:3 distinguish:2 lkhansen:1 activity:1 adapted:2 ahead:1 kronecker:1 pruned:2 wpq:7 designated:1 according:4 poor:1 disconnected:1 remain:1 slightly:1 rtrl:5 unity:1 shallow:1 cun:1 modification:2 lyngby:1 equation:1 discus:2 initiate:1 addison:1 fed:1 available:5 apply:1 observe:1 denker:1 appropriate:1 permanently:1 hampered:1 denotes:6 running:1 cf:1 include:1 newton:4 concatenated:1 build:1 quantity:1 damage:2 dependence:1 traditional:1 diagonal:1 gradient:2 morten:5 thank:2 topic:1 lthe:1 denmark:2 length:1 besides:1 index:5 ratio:4 providing:1 design:1 observation:2 benchmark:4 descent:2 excluding:1 varied:1 redwood:1 arbitrary:1 retrained:2 tentative:1 able:1 suggested:3 beyond:1 usually:1 dynamical:1 appeared:1 memory:8 omlin:1 difficulty:2 rely:1 natural:1 predicting:1 recursion:1 scheme:7 improve:1 morasso:1 numerous:1 dtu:1 woo:2 ironed:1 understanding:1 zh:1 fully:6 interesting:2 dop:1 s0:4 imposes:1 lo:1 obd:2 supported:1 allow:1 institute:2 slice:2 feedback:4 dimension:1 calculated:2 doesn:1 autoregressive:1 san:2 transaction:1 pruning:19 ons:1 francisco:2 zk:3 robust:1 ca:1 investigated:2 significance:1 main:1 icann:1 sunspot:7 slow:2 hassibi:5 experienced:1 clamped:1 remained:1 showing:1 dk:2 decay:16 dki:1 virtue:1 essential:1 burden:1 incorporating:1 importance:1 magnitude:3 nec:1 conditioned:1 likely:1 springer:1 nested:1 oct:1 marked:1 consequently:1 ann:1 huberman:1 wolff:1 svarer:6 pas:1 gauss:2 select:1 internal:9 accelerated:1 incorporate:1 |
7,229 | 988 | Comparing the prediction accuracy of
artificial neural networks and other
statistical models for breast cancer
survival
Harry B. Burke
Department of Medicine
New York Medical College
Valhalla, NY 10595
David B. Rosen
Department of Medicine
New York Medical College
Valhalla, NY 10595
Philip H. Goodman
Department of Medicine
University of Nevada School of Medicine
Reno, Nevada 89520
Abstract
The TNM staging system has been used since the early 1960's
to predict breast cancer patient outcome. In an attempt to increase prognostic accuracy, many putative prognostic factors have
been identified. Because the TNM stage model can not accommodate these new factors, the proliferation of factors in breast
cancer has lead to clinical confusion. What is required is a new
computerized prognostic system that can test putative prognostic
factors and integrate the predictive factors with the TNM variables in order to increase prognostic accuracy. Using the area under the curve of the receiver operating characteristic, we compare
the accuracy of the following predictive models in terms of five
year breast cancer-specific survival: pTNM staging system, principal component analysis, classification and regression trees, logistic
regression, cascade correlation neural network, conjugate gradient
descent neural, probabilistic neural network, and backpropagation
neural network. Several statistical models are significantly more ac-
1064
Harry B. Burke, David B. Rosen, Philip H. Goodman
curate than the TNM staging system. Logistic regression and the
backpropagation neural network are the most accurate prediction
models for predicting five year breast cancer-specific survival
1
INTRODUCTION
For over thirty years measuring cancer outcome has been based on the TNM staging
system (tumor size, number of lymph nodes with metastatic disease, and distant
metastases) (Beahr et. al., 1992). There are several problems with this model
(Burke and Henson , 1993). First, it is not very accurate, for breast cancer it is
44% accurate. Second its accuracy can not be improved because predictive variables can not be added to the model. Third, it does not apply to all cancers. In
this paper we compare computerized prediction models to determine if they can
improve prognostic accuracy. Artificial neural networks (ANN) are a class of nonlinear regression and discrimination models. ANNs are being used in many areas
of medicine, with several hundred articles published in the last year. Representative areas of research include anesthesiology (Westenskow et. al., 1992) , radiology
(Tourassi et. al. , 1992) , cardiology (Leong and Jabri, 1982) , psychiatry (Palombo,
1992), and neurology (Gabor and Seyal, 1992). ANNs are being used in cancer
research including image processing (Goldberg et. al., 1992) , analysis of laboratory data for breast cancer diagnosis (0 Leary et. al., 1992), and the discovery of
chemotherapeutic agents (Weinstein et . al., 1992). It should be pointed out that
the analyses in this paper rely upon previously collected prognostic factors. These
factors were selected for collection because they were significant in a generalized
linear model such as the linear or logistic models . There is no predictive model that
can improve upon linear or logistic prediction models when the predictor variables
meet the assumptions of these models and there are no interactions. Therefore
he objective of this paper is not to outperform linear or logistic models on these
data. Rather, our objective is to show that, with variables selected by generalized
linear models, artificial neural networks can perform as well as the best traditional
models . There is no a priori reason to believe that future prognostic factors will
be binary or linear, and that there will not be complex interactions between prognostic factors. A further objective of this paper is to demonstrate that artificial
neural networks are likely to outperform the conventional models when there are
unanticipated nonmonotonic factors or complex interactions.
2
2.1
METHODS
DATA
The Patient Care Evaluation (PCE) data set is collected by the Commission on
Cancer of the American College of Surgeons (ACS). The ACS, in October 1992,
requested cancer information from hospital tumor registries in the United States.
The ACS asked for the first 25 cases of breast cancer seen at that institution in 1983,
and it asked for follow up information on each of these 25 patients through the date
of the request. These are only cases of first breast cancer. Follow-up information
included known deaths. The PCE data set contains, at best, eight year follow-up.
Prediction Accuracy of Models for Breast Cancer Survival
1065
We chose to use a five year survival end-point. This analysis is for death due to
breast cancer, not all cause mortality.
For this analysis cases with missing data, and cases censored before five years, are
not included so that the prediction models can be compared without putting any
prediction model at a disadvantage. We randomly divided the data set into training,
hold-out, and testing subsets of 3,100, 2,069, and 3,102 cases, respectively.
2.2
MODELS
The TMN stage model used in this analysis is the pathologic model (pTNM) based
on the 1992 American Joint Committee on Cancer's Manual for the Staging of
Cancer (Beahr et. al., 1992). The pathologic model relies upon pathologically determined tumor size and lymph nodes, this contrasts with clinical staging which
relies upon the clinical examination to provide tumor size and lymph node information. To determine the overall accuracy of the TNM stage model we compared
the model's prediction for each patient, where the individual patient's prediction
is the fraction of all the patients in that stage who survive, to each patient's true
outcome.
Principal components analysis, is a data reduction technique based on the linear
combinations of predictor variables that minimizes the variance across patients (Jollie, 1982). The logistic regression analysis is performed in a stepwise manner, without interaction terms, using the statistical language S-PLUS (S-PLUS, 1992), with
the continuous variable age modeled with a restricted cubic spline to avoid assuming
linearity (Harrell et. al., 1988). Two types of Classification and Regression Tree
(CART) (Breiman et. al., 1984) analyses are performed using S-PLUS. The first
was a 9-node pruned tree (with 10-fold cross validation on the deviance), and the
second was a shrunk tree with 13.7 effective nodes.
The multilayer perceptron neural network training in this paper is based on the
maximum likelihood function unless otherwise stated, and backpropagation refers
to gradient descent. Two neural networks that are not multilayer perceptrons are
tested. They are the Fuzzy ARTMAP neural network (Carpenter et. al., 1991) and
the probabilistic neural network (Specht, 1990).
2.3
ACCURACY
The measure of comparative accuracy is the area under the curve of the receiver
operating characteristic (Az) . Generally, the Az is a nonparametric measure of
discrimination. Square error summarizes how close each patient's predicted value is
to its true outcome. The Az measures the relative goodness of the set of predictions
as a whole by comparing the predicted probability of each patient with that of all
other patients. The computational approach to the Az that employs the trapezoidal
approximation to the area under the receiver operating characteristic curve for
binary outcomes was first reported by Bamber (Bamber, 1975), and later in the
medical literature by Hanley (Hanley and McNeil, 1982). This was extended by
Harrell (Harrell et. al., 1988) to continuous outcomes.
Harry B. Burke, David B. Rosen, Philip H. Goodman
1066
Table 1: PCE 1983 Breast Cancer Data: 5 Year Survival Prediction, 54 Variables.
PREDICTION MODEL
pTNM Stages
Principal Components Analysis
CART, pruned
CART, shrunk
Stepwise Logistic regression
Fuzzy ARTMAP ANN
Cascade correlation ANN
Conjugate gradient descent ANN
Probabilistic ANN
Backpropagation ANN
ACCURACY?
SPECIFICATIONS
.720
.714
.753
.762
.776
.738
.761
.774
.777
.784
O,I,I1A,I1B,IIIA,I1IB,IV
one scaling iteration
9 nodes
13.7 nodes
with cubic splines
54-F2a, 128-1
54-21-1
54-30-1
bandwidth = 16s
54-5-1
* The area under the curve of the receiver operating characteristic.
3
RESULTS
All results are based on the independent variable sample not used for training (i.e.,
the testing data set), and all analyses employ the same testing data set. Using
the PCE breast cancer data set, we can assess the accuracy of several prediction
models using the most powerful of the predictor variables available in the data set
(See Table 1).
Principal components analysis is not expected to be a very accurate model; with
one scaling iteration, its accuracy is .714. Two types of classification and regression trees (CART), pruned and shrunk, demonstrate accuracies of .753 and .762,
respectively. Logistic regression with cubic splines for age has an accuracy of .776.
In addition to the backpropagation neural network and the probabilistic neural network, three types of neural networks are tested. Fuzzy ARTMAP's accuracy is
the poorest at .738. It was too computationally intensive to be a practical model.
Cascade-correlation and conjugate gradient descent have the potential to do as well
as backpropagation. The PNN accuracy is .777. The PNN has many interesting
features, but it also has several drawbacks including its storage requirements. The
backpropagation neural network's accuracy is .784.4.
4
DISCUSSION
For predicting five year breast cancer-specific survival, several computerized prediction models are more accurate than the TNM stage system, and artificial neural
networks are as good as the best traditional statistical models.
References
Bamber D (1975). The area above the ordinal dominance graph and the area below
the receiver operating characteristic. J Math Psych 12:387-415.
Beahrs OH, Henson DE, Hutter RVP, Kennedy BJ (1992). Manual for staging of
Prediction Accuracy of Models for Breast Cancer Survival
1067
cancer, 4th ed. Philadelphia: JB Lippincott.
Burke HB, Henson DE (1993) . Criteria for prognostic factors and for an enhanced
prognostic system. Cancer 72:3131-5.
Breiman L, Friedman JH, Olshen RA (1984). Classification and Regression Trees.
Pacific Grove, CA: Wadsworth and Brooks/Cole.
Carpenter GA, Grossberg S, Rosen DB (1991). Fuzzy ART: Fast stable learning
and categorization of analog patterns by an adaptive resonance system. Neural
Networks 4:759-77l.
Gabor AJ, M. Seyal M (1992) . Automated interictal EEG spike detection using
artificial neural networks. Electroencephalogr Clin Neurophysiology 83 :271-80.
Goldberg V, Manduca A, Ewert DL (1992). Improvement in specificity of ultrasonography for diagnosis of breast tumors by means of artificial intelligence. Med
Phys 19:1275-8l.
Hanley J A, McNeil BJ (1982). The meaning of the use of the area under the receiver
operating characteristic (ROC) curve. Radiology 143:29-36.
Harrell FE, Lee KL, Pollock BG (1988). Regression models in clinical studies:
determining relationships between predictors and response. J Natl Cancer Instit
80:1198-1202.
Jollife IT (1986). Principal Component Analysis. New York: Springer-Verlag, 1986.
Leong PH, J abri MA (1982). MATIC - an intracardiac tachycardia classification
system. PACE 15:1317-31,1982.
O'Leary TJ, Mikel UV, Becker RL (1992). Computer-assisted image interpretation:
use of a neural network to differentiate tubular carcinoma from sclerosing adenosis.
Modern Pathol 5:402-5.
Palombo SR (1992). Connectivity and condensation in dreaming. JAm Psychoanal
Assoc 40:1139-59.
S-PLUS (1991), v 3.0. Seattle, WA; Statistical Sciences, Inc.
Specht DF (1990). Probabilistic neural networks. Neural Networks 3:109-18.
Tourassi GD, Floyd CE, Sostman HD, Coleman RE (1993). Acute pulmonary
embolism: artificial neural network approach for diagnosis. Radiology 189:555-58.
Weinstein IN, Kohn KW, Grever MR et. al. (1992) Neural computing in cancer
drug development: predicting mechanism of action. Science 258:447-51.
Westenskow DR, Orr JA, Simon FH (1992) . Intelligent alarms reduce anesthesiologist's response time to critical faults. Anesthesiology 77:1074-9, 1992.
| 988 |@word neurophysiology:1 prognostic:11 accommodate:1 reduction:1 contains:1 united:1 comparing:2 distant:1 discrimination:2 intelligence:1 selected:2 coleman:1 institution:1 math:1 node:7 seyal:2 five:5 weinstein:2 manner:1 ra:1 expected:1 proliferation:1 linearity:1 what:1 minimizes:1 psych:1 fuzzy:4 assoc:1 medical:3 before:1 meet:1 chose:1 plus:4 grossberg:1 practical:1 thirty:1 testing:3 backpropagation:7 area:9 drug:1 cascade:3 significantly:1 gabor:2 deviance:1 refers:1 specificity:1 cardiology:1 tmn:1 close:1 ga:1 storage:1 conventional:1 missing:1 artmap:3 oh:1 hd:1 bamber:3 enhanced:1 goldberg:2 trapezoidal:1 disease:1 asked:2 f2a:1 lippincott:1 predictive:4 surgeon:1 upon:4 anesthesiologist:1 joint:1 dreaming:1 fast:1 effective:1 artificial:8 outcome:6 nonmonotonic:1 metastatic:1 otherwise:1 radiology:3 differentiate:1 nevada:2 interaction:4 date:1 az:4 seattle:1 requirement:1 comparative:1 categorization:1 ac:4 school:1 pathologic:2 predicted:2 tourassi:2 drawback:1 shrunk:3 jam:1 ja:1 assisted:1 burke:5 hold:1 predict:1 bj:2 early:1 valhalla:2 fh:1 cole:1 electroencephalogr:1 rather:1 avoid:1 breiman:2 improvement:1 likelihood:1 contrast:1 psychiatry:1 overall:1 classification:5 priori:1 development:1 resonance:1 art:1 wadsworth:1 abri:1 kw:1 survive:1 future:1 rosen:4 jb:1 metastasis:1 spline:3 intelligent:1 employ:2 modern:1 randomly:1 individual:1 attempt:1 friedman:1 detection:1 evaluation:1 natl:1 tj:1 staging:7 accurate:5 grove:1 censored:1 unless:1 tree:6 iv:1 re:1 hutter:1 disadvantage:1 measuring:1 goodness:1 subset:1 hundred:1 predictor:4 too:1 commission:1 reported:1 gd:1 probabilistic:5 lee:1 leary:2 connectivity:1 mortality:1 dr:1 american:2 potential:1 de:2 orr:1 harry:3 inc:1 ewert:1 bg:1 performed:2 later:1 i1a:1 curate:1 simon:1 ass:1 square:1 accuracy:19 variance:1 characteristic:6 who:1 computerized:3 kennedy:1 published:1 anns:2 phys:1 manual:2 ed:1 follow:3 response:2 improved:1 stage:6 correlation:3 nonlinear:1 logistic:8 aj:1 believe:1 true:2 laboratory:1 death:2 floyd:1 criterion:1 generalized:2 demonstrate:2 confusion:1 image:2 meaning:1 rl:1 analog:1 he:1 interpretation:1 significant:1 uv:1 pointed:1 language:1 henson:3 specification:1 stable:1 operating:6 acute:1 verlag:1 binary:2 fault:1 seen:1 care:1 mr:1 pathol:1 determine:2 clinical:4 cross:1 tubular:1 divided:1 prediction:15 regression:11 breast:16 patient:11 intracardiac:1 df:1 multilayer:2 iteration:2 pnn:2 addition:1 condensation:1 goodman:3 sr:1 cart:4 med:1 db:1 leong:2 hb:1 automated:1 identified:1 bandwidth:1 registry:1 rvp:1 reduce:1 intensive:1 kohn:1 becker:1 york:3 cause:1 action:1 generally:1 nonparametric:1 westenskow:2 ph:1 instit:1 outperform:2 carcinoma:1 pace:1 diagnosis:3 dominance:1 putting:1 ce:1 ptnm:3 graph:1 mcneil:2 fraction:1 year:9 powerful:1 putative:2 summarizes:1 scaling:2 poorest:1 interictal:1 fold:1 pruned:3 department:3 pacific:1 request:1 combination:1 conjugate:3 across:1 restricted:1 computationally:1 previously:1 committee:1 manduca:1 mechanism:1 ordinal:1 specht:2 end:1 available:1 apply:1 eight:1 include:1 clin:1 medicine:5 hanley:3 objective:3 added:1 spike:1 traditional:2 gradient:4 pathologically:1 ultrasonography:1 philip:3 collected:2 reason:1 mikel:1 assuming:1 modeled:1 relationship:1 october:1 olshen:1 fe:1 stated:1 perform:1 descent:4 embolism:1 extended:1 unanticipated:1 david:3 required:1 kl:1 lymph:3 brook:1 below:1 pattern:1 including:2 critical:1 rely:1 examination:1 predicting:3 improve:2 anesthesiology:2 reno:1 philadelphia:1 literature:1 discovery:1 determining:1 relative:1 interesting:1 age:2 validation:1 integrate:1 agent:1 article:1 cancer:26 last:1 iiia:1 harrell:4 jh:1 perceptron:1 curve:5 collection:1 adaptive:1 receiver:6 neurology:1 matic:1 continuous:2 table:2 ca:1 eeg:1 requested:1 complex:2 jabri:1 tachycardia:1 whole:1 alarm:1 carpenter:2 representative:1 roc:1 cubic:3 ny:2 third:1 pollock:1 tnm:7 specific:3 survival:8 dl:1 stepwise:2 beahr:2 likely:1 palombo:2 springer:1 pulmonary:1 relies:2 ma:1 ann:6 included:2 determined:1 principal:5 tumor:5 pce:4 hospital:1 perceptrons:1 college:3 tested:2 |
7,230 | 989 | Convergence Properties of the K-Means
Algorithms
Leon Bottou
Neuristique,
28 rue des Petites Ecuries,
75010 Paris, France
leonCneuristique.fr
Yoshua Bengio"
Dept. LR.O.
Universite de Montreal
Montreal, Qc H3C-3J7, Canada
bengioyCiro.umontreal.ca
Abstract
This paper studies the convergence properties of the well known
K-Means clustering algorithm. The K-Means algorithm can be described either as a gradient descent algorithm or by slightly extending the mathematics of the EM algorithm to this hard threshold
case. We show that the K-Means algorithm actually minimizes the
quantization error using the very fast Newton algorithm.
1
INTRODUCTION
K-Means is a popular clustering algorithm used in many applications, including the
initialization of more computationally expensive algorithms (Gaussian mixtures,
Radial Basis Functions, Learning Vector Quantization and some Hidden Markov
Models). The practice of this initialization procedure often gives the frustrating
feeling that K-Means performs most of the task in a small fraction of the overall
time. This motivated us to better understand this convergence speed.
A second reason lies in the traditional debate between hard threshold (e .g. KMeans, Viterbi Training) and soft threshold (e.g. Gaussian Mixtures, Baum Welch)
algorithms (Nowlan, 1991) . Soft threshold algorithms are often preferred because
they have an elegant probabilistic framework and a general optimization algorithm
named EM (expectation-maximization) (Dempster, Laird and Rubin, 1977). In the
case of a gaussian mixture, the EM algorithm has recently been shown to approximate the Newton optimization algorithm (Xu and Jordan, 1994). We prove in this
"also, AT&T Bell Labs, Holmdel, NJ 07733
586
Leon Bottou, Yoshua Bengio
paper that the corresponding hard threshold algorithm) K-Means) minimizes the
quantization error using exactly the Newton algorithm.
In the next section) we derive the K-Means algorithm as a gradient descent procedure. Section 3 extends the mathematics of the EM algorithm to the case of
K-Means . This second derivation of K-Means provides us with proper values for
the learning rates. In section 4 we show that this choice of learning rates optimally rescales the parameter space using Newton)s method. Finally) in section 5
we present and discuss a few experimental results comparing various versions of the
K-Means algorithm. The 5 clustering algorithms presented here were chosen for a
good coverage of the algorithms related to K-Means) but this paper does not have
the ambition of presenting a literature survey on the subject.
2
K-MEANS AS A GRADIENT DESCENT
Given a set of P examples (Xi) the K-Means algorithm computes k prototypes
= (Wk) which minimize the quantization error, i.e.) the average distance between
each pattern and the closest prototype:
W
E (W )
=
def ""'
L....J L ( Xi) W )
.
t
=
L....J -1 ml.n( Xi
. 2
t
def ""'
- Wk)
2
(1)
Writing Si (w) for the subscript of the closest prototype to example Xi, we have
(2)
2.1
GRADIENT DESCENT ALGORITHM
We can then derive a gradient descent algorithm for the quantization error:
~W = - ( t 8~;:). This leads to the following batch update equation (updating prototypes after presenting all the examples) :
(3)
We can also derive a corresponding online algorithm which updates the prototypes
after the presentation of each pattern Xi:
A
OL(Xi) w)
_
u.W -
-(t
OW
)
l.e.)
if k = Si(W)
otherwise.
(4)
The proper value of the learning rate (t remain to be specified in both batch and
online algorithms. Convergence proofs for both algorithms (Bottou) 1991) exist
for decreasing values of the learning rates satisfying the conditions E (t = 00 and
E (; < 00. Following (Kohonen) 1989), we could choose (t = (o/t . We prove
however in this paper that there exist a much better choice of learning rates.
Convergence Properties of the K-Means Algorithms
3
3.1
587
K-MEANS AS AN EM STYLE ALGORITHM
EM STYLE ALGORITHM
The following derivation of K-Means is similar to the derivation of (MacQueen,
1967). We insist however on the identity between this derivation and the mathematics of EM (Liporace, 1976) (Dempster, Laird and Rubin, 1977).
Although [{-Means does not fit in a probabilistic framework, this similarity holds
for a very deep reason: The semi-ring of probabilies (!R+, +, x) and the idempotent semi-ring of hard-threshold scores (!R, Min, +) share the most significant algebraic properties (Bacceli, Cohen and Olsder, 1992). This assertion completely
describes the similarities and the potential differences between soft-threshold and
hard-threshold algorithms. A complete discussion however stands outside the scope
of this paper.
The principle of EM is to introduce additional "hidden" variables whose knowledge
would make the optimization problem easier. Since these hidden variables are unknown, we maximize an auxiliary function which averages over the values of the
hidden variables given the values of the parameters at the previous iteration. In
our case, the hidden variables are .the assignments Si(W) of the patterns to the prototypes. Instead of considering the expected value over the distribution on these
hidden variables, we just consider the values of the hidden variables that minimize
the cost, given the previous values of the parameters:
'"' 1 (
Q( WI ,W ) def
= L...J
'2 Xi - WI &.(w) )2
i
The next step consists then in finding a new set of prototypes Wi which mllllmizes Q( Wi, w) where w is the previous set of prototypes. We can analytically
compute the explicit solution of this minimization problem. Solving the equation
8Q(w' , W)/8W'k = 0 yields:
1
= 111
'"'
(5)
L...J Xi
k i:k=&i(w)
where Nk is the number of examples assigned to prototype Wk . The algorithm
consists in repeatedly replacing w by Wi using update equation (6) until convergence.
Since si(w' ) is by definition the best assignment of patterns Xi to the prototypes
w~, we have the following inequality:
wkI
E(w' ) - Q(w' , w) =
~ L(Xi -
Wi &.(w,))2 - (Xi - Wi Si(W))2 ::; 0
i
Using this result, the identity E(w) = Q(w, w) and the definition of Wi, we can
derive the following inequality:
E(w' ) - E(w)
E(w' ) - Q(w' , w) + Q(w ' , w) - Q(w, w)
< Q(w' , w) - Q(w, w) ::; 0
Each iteration of the algorithm thus decreases the otherwise positive quantization
error E (equation 1) until the error reaches a fixed point where condition W*' = w*
is verified (unicity of the minimum of Q(., w*)). Since the assignment functions
Si (w) are discrete, there is an open neighborhood of w* on which the assignments
are constant. According to their definition, functions E(.) and Q(., w*) are equal
on this neighborhood. Being the minimum of function Q(.,w*), the fixed point w*
of this algorithm is also a local minimum of the quantization error E. 0
588
3.2
Leon Bottou, Yoshua Bengio
BATCH K-MEANS
The above algorithm (5) can be rewritten in a form similar to that of the batch
gradient descent algorithm (3).
=
if k
S(Xi' w)
otherwise.
(6)
This algorithm is thus equivalent to a batch gradient descent with a specific, prototype dependent, learning rate k ?
J
3.3
ONLINE K-MEANS
The online version of our EM style update equation (5) is based on the computation
of the mean flt of the examples Xl, ... ,Xt with the following recursive formula:
flt+l
= t~l (t flt + Xt+l) = flt + t~l (Xt+l -
fld
Let us introduce new variables nk which count the number of examples so far
assigned to prototype Wk. We can then rewrite (5) as an online update applied
after the presentation of each pattern Xi:
D..nk
D..wk
{
{ n\
=
1 if k S(Xi' w)
0 otherwise.
(Xi - Wk)
0
if k = S(Xi'W)
otherwise.
(7)
This algorithm is equivalent to an online gradient descent (4) with a specific, prototype dependent, learning rate nlk. Unlike in the batch case, the pattern assignments
S(Xi , w) are thus changing after each pattern presentation. Before applying this algorithm, we must of course set nk to zero and Wk to some initial value. Various
methods have been proposed including initializing Wk with the first k patterns.
3.4
CONVERGENCE
General convergence proofs for the batch and online gradient descent (Bottou, 1991;
Driancourt, 1994) directly apply for all four algorithms. Although the derivatives
are undefined on a few points, these theorems prove that the algorithms almost
surely converge to a local minimum because the local variations of the loss function
are conveniently bounded (semi-differentiability). Unlike previous results, the above
convergence proofs allow for non-linearity, non-differentiability (on a few points)
(Bottou, 1991), and replacing learning rates by a positive definite matrix (Driancourt, 1994).
4
K-MEANS AS A NEWTON OPTIMIZATION
We prove in this section that Batch K-Means (6) applies the Newton algorithm.
4.1
THE HESSIAN OF K-MEANS
Let us compute the Hessian H of the K-Means cost function (2). This matrix
contains the second derivatives of the cost E(w) with respect to each pair of parameters. Since E(w) is a sum of terms L(Xi'W), we can decompose H as the sum
589
Convergence Properties of the K-Means Algorithms
of matrices Hi for each term of the cost function:
L(Xi'W)
= ~in~(xi -
Wk)2.
Furthermore, the Hi can be decomposed in blocks corresponding to each pair of
prototypes. Since L( Xi, w) depends only on the closest prototype to pattern Xi, all
these blocks are zero except block (Si (w), Si (w)) which is the identity matrix. Summing the partial Hessian matrices Hi thus gives a diagonal matrix whose diagonal
elements are the counts of examples Nk assigned to each prototype.
H=(T
o
We can thus write the Newton update of the parameters as follows:
~w = _H- 1 oE(w)
ow
which can be exactly rewritten as the batch EM style algorithm (6) presented earlier:
~w
k
4.2
="
{ 0~k (Xi - Wk)
L..J
i
=
if k ~(Xi' w)
otherwlse.
(8)
CONVERGENCE SPEED
When optimizing a quadratic function, Newton's algorithm requires only one step .
In the case of a non quadratic function, Newton's algorithm is superlinear if we
can bound the variations of the second derivatives. Standard theorems that bound
this variation using the third derivative are not useful for K-Means because the
gradient of the cost function is discontinuous. We could notice that the variations
of the second derivatives are however nicely bounded and derive similar proofs for
K-Means.
For the sake of brevity however, we are just giving here an intuitive argument: we
can make the cost function indefinitely differentiable by rounding up the angles
around the non differentiable points. We can even restrict this cost function change
to an arbitrary small region of the space. The iterations of K-Means will avoid
this region with a probability arbitrarily close to 1. In practice, we obtain thus a
superlinear convergence.
Batch K-Means thus searches for the optimal prototypes at Newton speed. Once
it comes close enough to the optimal prototypes (i .e. the pattern assignment is
optimal and the cost function becomes quadratic), K-Means jumps to the optimum
and terminates.
Online K-Means benefits of these optimal learning rates because they remove the
usual conditioning problems of the optimization. However , the stochastic noise
induced by the online procedure limits the final convergence of the algorithm. Final
convergence speed is thus essentially determined by the schedule of the learning
rates.
Online K-Means also benefits from the redundancies of the training set. It converges
significantly faster than batch K-Means during the first training epochs (Darken
Leon Bottou, Yoshua Bengio
590
IaI
BII Co.t
Co.t
2500
2'00
2300
2200
2100
2000
1900
1800
1700
1600
1500
1600
1300
1200
1100
1000
900
800
700
600
500
'00
300
200
100
1
2
3
,
5
6
7
8
9 10 11 12 13 16 15 16 17 18 19 20
Figure 1: E t - Eoo versus t. black circles: online K-Means; black squares: batch
K-Means; empty circles: online gradient; empty squares: batch gradient; no mark:
EM +Gaussian mixture
and Moody, 1991). After going through the first few patterns (depending of the
amount of redundancy), online K-Means indeed improves the prototypes as much
as a complete batch K-Means epoch. Other researchers have compared batch and
online algorithms for neural networks, with similar conclusions (Bengio, 1991).
5
EXPERIMENTS
Experiments have been carried out with Fisher's iris data set, which is composed of
150 points in a four dimensional space representing physical measurements on various species of iris flowers. Codebooks of six prototypes have been computed using
both batch and online K-Means with the proper learning rates (6) and (7). These
results are compared with those obtained using both gradient descent algorithms
(3) and (4) using learning rate ft = 0.03/t that we have found optimal. Results are
also compared with likelihood maximization with the EM algorithm, applied to a
mixture of six Gaussians, with fixed and uniform mixture weights, and fixed unit
variance. Inputs were scaled down empirically so that the average cluster variance
was around unity. Thus only the cluster positions were learned, as for the K-Means
algorithms.
Each run of an algorithm consists in (a) selecting a random initial set of prototypes,
(b) running the algorithm during 20 epochs and recording the error measure E t after
each epoch, (c) running the batch K-Means algorithm! during 40 more epochs in
order to locate the local minimum Eoo corresponding to the current initialization
of the algorithm. For the four K-Means algorithms, E t is the quantization error
(equation 1). For the Gaussian mixture trained with EM, the cost E t is the negative
1 except
for the case of the mixture of Gaussians, in which the EM algorithm was applied
Convergence Properties of the K-Means Algorithms
1
2
3
?
5
6
7
8
9
10 11 12 13 U
591
15 16 17 18 19
Figure 2: E~-;~~~oo versus t . black circles: online K-Means; black squares: batch
K-Means ; empty circles: online gradient; empty squares: batch gradient ; no mark:
EM+Gaussian mixture
logarithm of the likelihood of the data given the model.
Twenty trials were run for each algorithm. Using more than twenty runs did not
improve the standard deviation of the averaged measures because various initializations lead to very different local minima. The value Eoo of the quantization error
on the local minima ranges between 3300 and 5800. This variability is caused by
the different initializations and not by the different algorithms. The average values
of Eoo for each algorithm indeed fall in a very small range (4050 to 4080).
Figure 1 shows the average value of the residual error E t - Eoo during the first
20 epochs. Online K-Means (black circles) outperforms all other algorithms during
the first five epochs and stabilizes on a level related to the stochastic noise of the
online procedure. Batch K-Means (black squares) initially converges more slowly
but outperforms all other methods after 5 epochs. All 20 runs converged before the
15th epoch. Both gradients algorithms display poor convergence because they do
not benefit of the Newton effect. Again, the online version (white circles) starts
faster then the batch version (white square) but is outperformed in the long run.
The negative logarithm of the Gaussian mixture is shown on the curve with no
point marks , and the scale is displayed on the right of Figure 1.
Figure 2 show the final convergence properties of all five algorithms. The evolutions
of the ratio (Et+l - Eoo)/(Et - Eoo) characterize the relative improvement of the
residual error after each iteration. All algorithms exhibit the same behavior after a
few epochs except batch K-Means (black squares). The fast convergence ofthis ratio
to zero demonstrates the final convergence of batch K-Means. The EM algorithm
displays a better behavior than all the other algorithms except batch K-Means.
Clearly, however , its relative improvement ratio doesn't display the fast convergence
behavior of batch K-Means.
Uon Bottou, Yoshua Bengio
592
The online K-Means curve crosses the batch K-Means curve during the second
epoch, suggesting that it is better to run the online algorithm (7) during one epoch
and then switch to the batch algorithm (6).
6
CONCLUSION
We have shown with theoretical arguments and simple experiments that a well
implemented K-Means algorithm minimizes the quantization error using Newton's
algorithm. The EM style derivation of K-Means shows that the mathematics of EM
are valid well outside the framework of probabilistic models. Moreover the provable
convergence properties of the hard threshold K-Means algorithm are superior to
those of the EM algorithm for an equivalent soft threshold mixture of Gaussians.
Extending these results to other hard threshold algorithms (e .g. Viterbi Training)
is an interesting open question.
References
Bacceli, F., Cohen, G., and Olsder, G. J. (1992). Synchronization and Linearity.
Wiley.
Bengio, Y. (1991). Artificial Neural Networks and their Application to Sequence
Recognition. PhD thesis, McGill University, (Computer Science), Montreal,
Qc., Canada.
Bottou, 1. (1991). Une approche theorique de l'apprentissage connexioniste; applications a la reconnaissance de la parole. PhD thesis, Universite de Paris
XI.
Darken, C. and Moody, J. (1991). Note on learning rate schedules for stochastic
optimization. In Lippman, R. P., Moody, R., and Touretzky, D. S., editors,
Advances in Neural Information Processing Systems 3, pages 832-838, Denver,
CO. Morgan Kaufmann, Palo Alto.
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum-likelihood from
incomplete data via the EM algorithm. Journal of Royal Statistical Society B,
39:1-38.
Driancourt, X. (1994). Optimisation par descente de gradient stochastique ... , PhD
thesis, Universite de Paris XI, 91405 Orsay cedex, France.
Kohonen, T. (1989). Self-Organization and Associative Memory. Springer-Verlag,
Berlin, 3 edition.
Liporace, L. A . (1976). PTAH on continuous multivariate functions of Markov
chains. Technical Report 80193, Institute for Defense Analysis, Communication
Research Department.
MacQueen, J. (1967). Some methods for classification and analysis of multivariate
observations. In Proceedings of the Fifth Berkeley Symposium on Mathematics,
Statistics and Probability, Vol. 1, pages 281-296.
Nowlan, S. J. (1991). Soft Competitive Adaptation: Neural Network Learning Algorithms based on Fitting Statistical Mixtures. CMU-CS-91-126, School of Computer Science, Carnegie Mellon University, Pittsburgh, PA.
Xu, L. and Jordan, M. (1994). Theoretical and experimental studies of convergence properties of the em algorithm for unsupervised learning based on finite
mixtures. Presented at the Neural Networks for Computing Conference.
| 989 |@word trial:1 version:4 open:2 uon:1 initial:2 contains:1 score:1 selecting:1 outperforms:2 current:1 comparing:1 nowlan:2 si:8 must:1 remove:1 update:6 une:1 lr:1 indefinitely:1 provides:1 five:2 symposium:1 prove:4 consists:3 fitting:1 fld:1 introduce:2 indeed:2 expected:1 behavior:3 ol:1 insist:1 decreasing:1 decomposed:1 considering:1 becomes:1 wki:1 bounded:2 linearity:2 moreover:1 alto:1 minimizes:3 finding:1 nj:1 berkeley:1 neuristique:1 exactly:2 scaled:1 demonstrates:1 unit:1 positive:2 before:2 local:6 limit:1 subscript:1 black:7 initialization:5 co:3 range:2 averaged:1 practice:2 recursive:1 definite:1 block:3 lippman:1 procedure:4 bell:1 significantly:1 radial:1 superlinear:2 close:2 liporace:2 applying:1 writing:1 equivalent:3 baum:1 survey:1 welch:1 qc:2 variation:4 mcgill:1 pa:1 element:1 expensive:1 satisfying:1 updating:1 recognition:1 ft:1 initializing:1 region:2 oe:1 decrease:1 dempster:3 trained:1 solving:1 rewrite:1 basis:1 completely:1 various:4 derivation:5 fast:3 artificial:1 outside:2 neighborhood:2 whose:2 otherwise:5 statistic:1 h3c:1 laird:3 final:4 online:22 associative:1 sequence:1 differentiable:2 fr:1 adaptation:1 kohonen:2 intuitive:1 convergence:22 empty:4 optimum:1 extending:2 cluster:2 ring:2 converges:2 derive:5 depending:1 montreal:3 oo:1 school:1 coverage:1 auxiliary:1 implemented:1 come:1 c:1 discontinuous:1 stochastic:3 eoo:7 decompose:1 hold:1 around:2 viterbi:2 scope:1 stabilizes:1 outperformed:1 palo:1 minimization:1 clearly:1 j7:1 gaussian:7 avoid:1 improvement:2 likelihood:3 dependent:2 initially:1 hidden:7 going:1 france:2 overall:1 classification:1 equal:1 once:1 nicely:1 unsupervised:1 yoshua:5 report:1 few:5 composed:1 organization:1 mixture:13 undefined:1 chain:1 partial:1 incomplete:1 logarithm:2 circle:6 theoretical:2 soft:5 earlier:1 assertion:1 assignment:6 maximization:2 cost:9 deviation:1 frustrating:1 uniform:1 rounding:1 optimally:1 characterize:1 probabilistic:3 reconnaissance:1 moody:3 again:1 thesis:3 choose:1 slowly:1 derivative:5 style:5 suggesting:1 potential:1 de:7 wk:10 rescales:1 caused:1 depends:1 lab:1 approche:1 start:1 competitive:1 minimize:2 otherwlse:1 square:7 variance:2 kaufmann:1 yield:1 researcher:1 converged:1 reach:1 touretzky:1 definition:3 universite:3 proof:4 popular:1 knowledge:1 improves:1 schedule:2 actually:1 iai:1 furthermore:1 just:2 until:2 replacing:2 effect:1 evolution:1 analytically:1 assigned:3 white:2 during:7 self:1 iris:2 presenting:2 complete:2 performs:1 recently:1 umontreal:1 superior:1 physical:1 empirically:1 denver:1 cohen:2 conditioning:1 significant:1 measurement:1 mellon:1 mathematics:5 similarity:2 closest:3 multivariate:2 optimizing:1 verlag:1 inequality:2 arbitrarily:1 morgan:1 minimum:7 additional:1 surely:1 converge:1 maximize:1 semi:3 technical:1 faster:2 cross:1 long:1 ambition:1 essentially:1 expectation:1 optimisation:1 cmu:1 iteration:4 unlike:2 cedex:1 subject:1 induced:1 elegant:1 recording:1 jordan:2 orsay:1 bengio:7 enough:1 switch:1 fit:1 restrict:1 codebooks:1 prototype:21 motivated:1 six:2 defense:1 algebraic:1 hessian:3 repeatedly:1 deep:1 useful:1 amount:1 differentiability:2 exist:2 notice:1 discrete:1 write:1 carnegie:1 vol:1 redundancy:2 four:3 threshold:11 changing:1 verified:1 fraction:1 sum:2 run:6 angle:1 named:1 extends:1 almost:1 holmdel:1 def:3 hi:3 bound:2 display:3 quadratic:3 sake:1 speed:4 argument:2 min:1 leon:4 department:1 according:1 poor:1 remain:1 slightly:1 em:21 describes:1 terminates:1 wi:8 unity:1 computationally:1 equation:6 discus:1 count:2 gaussians:3 rewritten:2 apply:1 bii:1 batch:27 clustering:3 running:2 newton:12 giving:1 society:1 question:1 usual:1 traditional:1 diagonal:2 exhibit:1 gradient:17 ow:2 distance:1 berlin:1 reason:2 provable:1 parole:1 ratio:3 debate:1 negative:2 proper:3 unknown:1 twenty:2 observation:1 darken:2 markov:2 macqueen:2 finite:1 descent:10 displayed:1 variability:1 communication:1 locate:1 arbitrary:1 canada:2 pair:2 paris:3 specified:1 unicity:1 learned:1 flower:1 pattern:11 including:2 royal:1 memory:1 residual:2 representing:1 improve:1 idempotent:1 carried:1 epoch:12 literature:1 relative:2 synchronization:1 loss:1 par:1 interesting:1 versus:2 olsder:2 apprentissage:1 rubin:3 principle:1 editor:1 share:1 course:1 allow:1 understand:1 institute:1 fall:1 fifth:1 benefit:3 curve:3 stand:1 valid:1 computes:1 doesn:1 jump:1 feeling:1 far:1 approximate:1 preferred:1 ml:1 summing:1 pittsburgh:1 xi:26 search:1 continuous:1 ca:1 bottou:9 rue:1 did:1 noise:2 edition:1 xu:2 wiley:1 position:1 explicit:1 xl:1 lie:1 third:1 formula:1 theorem:2 down:1 specific:2 xt:3 flt:4 ofthis:1 quantization:10 phd:3 nk:5 easier:1 conveniently:1 applies:1 springer:1 nlk:1 identity:3 presentation:3 kmeans:1 fisher:1 hard:7 change:1 determined:1 except:4 specie:1 experimental:2 la:2 mark:3 brevity:1 dept:1 |
7,231 | 99 | 695
ANALOG IMPLEMENTATION OF SHUNTING
NEURAL NETWORKS
Bahram Nabet, Robert B. Darling, and Robert B. Pinter
Department of Electrical Engineering, FT-lO
University of Washington
Seattle, WA 98195
ABSTRACT
An extremely compact, all analog and fully parallel implementation of a class of shunting recurrent neural networks that is applicable to a wide variety of FET-based integration technologies is
proposed. While the contrast enhancement, data compression, and
adaptation to mean input intensity capabilities of the network are
well suited for processing of sensory information or feature extraction for a content addressable memory (CAM) system, the network
also admits a global Liapunov function and can thus achieve stable
CAM storage itself. In addition the model can readily function as
a front-end processor to an analog adaptive resonance circuit.
INTRODUCTION
Shunting neural networks are networks in which multiplicative, or shunting, terms
of the form Xi Lj f;(Xj) or Xi Lj Ij appear in the short term memory equations,
where Xi is activity of a cell or a cell population or an iso-potential portion of a
cell and Ii are external inputs arriving at each site. The first case shows recurrent
activity, while the second case is non-recurrent or feed forward. The polarity of
these terms signify excitatory or inhibitory interactions.
Shunting network equations can be derived from various sources such as the passive
membrane equation with synaptic interaction (Grossberg 1973, Pinter 1983), models
of dendritic interaction (RaIl 1977), or experiments on motoneurons (Ellias and
Grossberg 1975).
While the exact mechanisms of synaptic interactions are not known in every individual case, neurobiological evidence of shunting interactions appear in several
696
Nabet, Darling and Pinter
areas such as sensory systems, cerebellum, neocortex, and hippocampus (Grossberg
1973, Pinter 1987). In addition to neurobiology, these networks have been used to
successfully explain data from disciplines ranging from population biology (Lotka
1956) to psychophysics and behavioral psychology (Grossberg 1983).
Shunting nets have important advantages over additive models which lack the extra nonlinearity introduced by the multiplicative terms. For example, the total
activity of the network, shown by Li Xi, approaches a constant even as the input
strength grows without bound. This normalization in addition to being computationally desirable has interesting ramifications in visual psychophysics (Grossberg
1983). Introduction of multiplicative terms also provides a negative feedback loop
which automatically controls the gain of each cell, contributes to the stability of the
network, and allows for large dynamic range of the input to be processed by the
network. The automatic gain control property in conjunction with properly chosen
nonlinearities in the feedback loop makes the network sensitive to small input values
by suppressing noise while not saturating at high input values (Grossberg 1973).
Finally, shunting nets have been shown to account for short term adaptation to
input properties, such as adaptation level tuning and the shift of sensitivity with
background strength (Grossberg 1983), dependence of visual size preference and
latency of response on contrast and mean luminance, and dependence of temporal
and spatial frequency tuning on contrast and mean luminance (Pinter 1985).
IMPLEMENTATION
The advantages, generality, and applicability of shunting nets as cited in the previous section make their implementation very desirable, but digital implementation
of these networks is very inefficient due to the need for analog to digital conversion, multiplication and addition instructions, and implementation of iterative algorithms. A linear feedback class of these networks (Xi Lj !; (Xj)
Xi Li J{ijXj),
however, can be implemented very efficiently with simple, completely parallel and
all analog circuits.
=
FRAMEWORK
Figure 1 shows the design framework for analog implementation of a class of shunting nets. In this design addition (subtraction) is achieved, via Kirchoff's current
law by placing transistors in upper (lower) rails, and through the choice of depletion or enhancement mode devices. Multiplicative, or shunting, interconnections
are done by one transistor per interconnect, using a field-effect transistor (FET) in
the voltage-variable conductance region. Temporal properties are characterized by
cell membrane capacitance C, which can be removed, or in effect replaced by the
parasitic device capacitances, if higher speed is desired. A buffer stage is necessary
for correct polarity of interconnections and the large fan-out associated with high
connectivity of neural networks.
Analog Implementation of Shunting Neural Networks
Vdd
I.,
+
x.,
c
-t"
'.J
.
x?
J
Vss
Figure 1. Design framework for implementation of one cell in a
shunting network. Voltage output of other cells is connected to the
gate of transistors Qi,i'
Such a circuit is capable of implementing the general network equation:
(1)
Excitatory and inhibitory input current sources can also be shunted, with extra
circuitry, to implement non-recurrent shunting networks.
NMOS, CMOS and GALLIUM ARSENIDE
Since the basic cell of Fig. 1 is very similar to a standard logic gate inverter, but with
the transistors sized by gate width-to-Iength ratio to operate in the nonsaturated
current region, this design is applicable to a variety of FET technologies including
NMOS, CMOS, and gallium arsenide (GaAs).
A circuit made of all depletion-mode devices such as GaAs MESFET buffered FET
logic, can implement all the terms of Eq. (1) except shunting excitatory terms and
requires a level shifter in the buffer stage. A design with all enhancement mode
devices such as silicon NMOS can do the same but without a level shifter. With
the addition of p-channel devices, e.g. Si CMOS, all polarities and all terms of Eq.
(1) can be realized. As mentioned previously a buffer stage is necessary for correct
polarity of interconnections and fan out/fan in capacity.
Figure 2 shows a GaAs MESFET implementation with only depletion mode devices
which employs a level shifter as the buffer stage.
697
698
Nabet, Darling and Pinter
VDD-------------r------~--------------~--
INPUTS:
EXTERNAL OR FROM
PREVJOUS LAYER
EXCITATORY
CONNECTIONS
INHIBITORY
CONNECTIONS
GN~-L--~------~------~--
TUNABLE
SELF-RELAXATION
CONNECTION
OUTPUT TO
NEXT LAYER
VSS--~--....L....?
LEVEL SHIFT AND BUFFER STAGE
Figure 2. Gallium arsenide MESFET implementation with level
shifter and depletion mode devices. Lower rail transistors produce
shunting off-surround terms. Upper transistors can produce additive excitatory connections.
SPECIFIC IMPLEMENTATION
The simplest shunting network that can be implemented by the general framework
of Fig.1 is Fig. 2 with only inhibitory connections (lower rail transistors). This
circuit implements the network model
dX?
d/
= Ii -
a,X,
+ Xi(J(iXi) -
" J(ijXj)
Xi(L....J
j#i
(2),
The simplicity of the implementation is notable; a linear array with nearest neighbor
interconnects consists of only 5 transistors, 1-3 diodes, and if required 1 capacitor
per cell.
A discrete element version of this implementation has been constructed and shows
good agreement with expected properties. Steady state output is proportional to
the square root of a uniform input thereby compressing the input data and showing
adaptation to mean input intensity (figure 3). The network exhibits contrast enhancement of spatial edges which increases with higher mean input strength (figure
4). A point source input elicits an on-center off-surround response, similar to the
difference-of-Gaussians receptive field of many excitable cells. This 'receptive field'
becomes more pronounced as the input intensity increases, showing the dependence
of spatial frequency tuning on mean input level (figure 5). The temporal response
of the network is also input dependent since the time constant of the exponential
Analog Implementation of Shunting Neural Networks
decay of the impulse response decreases with input intensity. Finally, the dependence of the above properties on mean input strength can be tuned by varying the
conductance of the central FET.
700.0
---
600.0
500.0 ?
>
E
11.1
II
400.0
!:i
~
5
300.0
~
o
200.0
100.0
0.1
0.3
0.5
0 .7
0.9
1. 1
1.3
1.5
1.7
1.11
2.1
INPUT CURRENT rnA
Figure 3. Response of network to uniform input. Output is proportional to the square root of the input.
DEPENDENCE OF ENHANCEMENT ON MEAN INPUT
1.5
1.4
~
:J
0
...
1.3
a
1.2
i
1.1
N
II:
0
z
5
~
0
a
1.0
0 .11
f
...::l
0.8
IL
~
0 .7
0.11
2
5
3
7
CELL NUMBER
Figure 4. Response of network to spatial edge patterns with the
same contrast but increasing mean input level.
699
700
Nabet, Darling and Pinter
Imox
~
2 .0J rnA, llmox - J75.74 mil
1.0
0 .11
~
g
:>
0.11
0
Q
~
0 .7
!5
II.
I
iii
0.11
II:
0.5
i
0
Z
0.4
O.J
2
J
4
II
5
7
cEll NUMBER
INPUT
-x-
OUTPUT
Figure 5. Response of network to a point source input. Inset
shows the receptive field of fly's lamina monopolar cells (LMC of
Lucilia sericata). Horizontal axis of inset in visual angle, vertical
a.xis relative voltage units of hyperpolarization. Inset from Pinter
et al. (in preparation)
CONTENT ADDRESSABILITY AND RELATION TO ART
Using a theorem by Cohen and Grossberg (1983), it can be shown that the network
equa.tion (2) a.dmits the global Liapunov function
n
V --
-
"'(l?ln(xi)
~
1
>. - a'x
1 1' + K'x~)
1 1
n
+.!2 '"
;=1
~
K??x
IJ J'XL
.. ,
(3)
j,k=l
=
where>. is a constant, under the constraints Kij
Kji and Xi > O. This shows that
in response to an arbitrary input the network always approaches an equilibrium
point. The equilibria represent stored patterns and this is Content Addressable
Memory (CAM) property.
In addition, Eq. (2) is a special case of the feature representation field of an analog
adaptive resonance theory ART-2 circuit, (Carpenter and Grossberg 1987), and
hence this design can operate as a module in a learning multilayer ART architecture.
Analog Implementation of Shunting Neural Networks
FUTURE PLANS
Due to the very small number of circuit components required to construct a cell,
this implementation is quite adaptable to very high integration densities. A solid
state implementation of the circuit of figure (2) on a gallium arsenide substrate,
chosen for its superiority for opto-electronics applications, is in progress. The chip
includes monolithically fabricated photosensors for processing of visual information.
All of the basic components of the circuit have been fabricated and tested. With
standard 2 micron GaAs BFL design rules, a chip could contain over 1000 cells per
cm 2 , assuming an average of 20 inputs per cell.
CONCLUSIONS
?
?
?
?
?
?
The present work has the following distinguishing features:
Implements a mathematically well described and stable model.
Proposes a framework for implementation of shunting nets which are biologically
feasible, explain variety of psychophysical and psychological data and have many
desirable computational properties.
Has self-sufficient computational capabilities; especially suited for processing of sensory information in general and visual information in particular (N abet and Darling
1988).
Produces a 'good representation' of the input data which is also compatible with
the self-organizing multilayer neural network architecture ART-2.
Is suitable for implementation in variety of technologies.
Is parallel, analog, and has very little overhead circuitry.
..
-.
701
702
Nabet, Darling and Pinter
REFERENCES
Carpenter, G.A. and Grossberg, S. (1987) "ART 2: self organization of stable category recognition codes for analog input patterns,". Applied Optics 26, pp. 49194930.
Cohen,M.A. and Grossberg, S. (1983) "Absolute stability of global pattern formation and parallel memory storage by competitive neural networks" , IEEE Transactions on Systems Man and Cybernetics SMC-13, pp. 815-826.
Ellias, S.A. and Grossberg, S. (1975) "Pattern formation, contrast control, and
oscillations in the short term memory of shunting on-center off-surround networks"
Biological Cybernetics, 20, pp. 69-98.
Grossberg, S. (1973), "Contour enhancement, Short term memory and constancies
in reverberating neural networks," Studies in Applied Mathematics, 52, pp. 217257.
Grossberg, S. (1983), "The quantized geometry of visual space: the coherent computation of depth, form, and lightness." The behavioral and brain sciences, 6, pp.
625-692.
Lotka, A.J. (1956). Elements of mathematical biology. New York: Dover.
Nabet, B. and Darling, R.B. (1988). "Implementation of optical sensory neural networks with simple discrete and monolithic circuits," (Abstract) Neural Networks,
Vol.l, Suppl. 1, 1988, pp. 396.
Pinter, R.B., (1983). "The electrophysiological bases for linear and nonlinear product term lateral inhibition and the consequences for wide-field textured stimuli"
J. Theor. Bioi. 105 pp. 233-243.
Pinter, R.B. (1985) " Adaptation of spatial modulation transfer functions via nonlinear lateral inhibition" Bioi. Cybernetics 51, pp. 285-291.
Pinter, R.B. (1987) "Visual system neural networks: Feedback and feedforward lateral inhibition" Systems and Control Encyclopedia (ed.M.G. Singh) Oxford: Pergamon Press. pp. 5060-5065.
Pinter, R.B., Osorio, D., and Srinivasan, M.V., (in preperation) "Shift of edge
preference to scototaxis depends on mean luminance and is predicted by a matched
filter hypothesis in fly lamina cells"
RaIl, W. (1977). "Core conductor theory and cable properties of neurons" in Handbook of Physiology: The Nervous System vol. I, part I, Ed. E.R. Kandel pp. 39-97.
Bethesda, MD: American Physiological Society.
| 99 |@word version:1 compression:1 hippocampus:1 instruction:1 thereby:1 solid:1 electronics:1 tuned:1 suppressing:1 current:4 si:1 dx:1 readily:1 additive:2 nonsaturated:1 v:2 device:7 nervous:1 liapunov:2 iso:1 dover:1 short:4 core:1 provides:1 quantized:1 preference:2 mathematical:1 constructed:1 consists:1 overhead:1 behavioral:2 expected:1 monopolar:1 brain:1 automatically:1 little:1 increasing:1 becomes:1 matched:1 circuit:10 cm:1 fabricated:2 temporal:3 every:1 control:4 unit:1 appear:2 superiority:1 engineering:1 monolithic:1 consequence:1 oxford:1 modulation:1 gaas:4 smc:1 range:1 equa:1 grossberg:14 implement:4 addressable:2 area:1 physiology:1 storage:2 center:2 simplicity:1 rule:1 array:1 population:2 stability:2 exact:1 substrate:1 ixi:1 distinguishing:1 hypothesis:1 agreement:1 element:2 recognition:1 ft:1 module:1 fly:2 constancy:1 electrical:1 region:2 compressing:1 connected:1 decrease:1 removed:1 mentioned:1 cam:3 dynamic:1 vdd:2 singh:1 completely:1 textured:1 chip:2 various:1 formation:2 quite:1 interconnection:3 itself:1 advantage:2 transistor:9 net:5 interaction:5 product:1 adaptation:5 loop:2 ramification:1 organizing:1 achieve:1 pronounced:1 seattle:1 enhancement:6 produce:3 cmos:3 lamina:2 recurrent:4 nearest:1 ij:2 progress:1 eq:3 implemented:2 diode:1 predicted:1 correct:2 filter:1 implementing:1 dendritic:1 biological:1 theor:1 mathematically:1 equilibrium:2 kirchoff:1 circuitry:2 inverter:1 applicable:2 sensitive:1 successfully:1 rna:2 always:1 varying:1 voltage:3 mil:1 monolithically:1 conjunction:1 derived:1 properly:1 contrast:6 dependent:1 interconnect:1 lj:3 relation:1 proposes:1 resonance:2 spatial:5 integration:2 psychophysics:2 art:5 special:1 field:6 plan:1 construct:1 extraction:1 washington:1 biology:2 placing:1 future:1 stimulus:1 employ:1 individual:1 replaced:1 geometry:1 conductance:2 organization:1 edge:3 capable:1 necessary:2 desired:1 psychological:1 kij:1 gn:1 applicability:1 uniform:2 front:1 stored:1 cited:1 density:1 sensitivity:1 off:3 discipline:1 shunted:1 connectivity:1 central:1 external:2 american:1 inefficient:1 li:2 account:1 potential:1 nonlinearities:1 includes:1 notable:1 depends:1 multiplicative:4 root:2 tion:1 portion:1 competitive:1 capability:2 parallel:4 square:2 il:1 efficiently:1 cybernetics:3 processor:1 explain:2 synaptic:2 ed:2 frequency:2 pp:10 associated:1 gain:2 tunable:1 electrophysiological:1 adaptable:1 feed:1 higher:2 response:8 photosensors:1 done:1 generality:1 stage:5 horizontal:1 nonlinear:2 lack:1 mode:5 impulse:1 grows:1 effect:2 contain:1 hence:1 cerebellum:1 lmc:1 width:1 self:4 nabet:6 steady:1 fet:5 passive:1 ranging:1 hyperpolarization:1 cohen:2 analog:12 buffered:1 silicon:1 surround:3 automatic:1 tuning:3 mathematics:1 nonlinearity:1 stable:3 inhibition:3 base:1 buffer:5 motoneuron:1 subtraction:1 ii:7 desirable:3 characterized:1 shunting:21 qi:1 basic:2 multilayer:2 normalization:1 represent:1 suppl:1 achieved:1 cell:17 background:1 addition:7 signify:1 source:4 extra:2 operate:2 bfl:1 capacitor:1 feedforward:1 iii:1 variety:4 xj:2 psychology:1 architecture:2 shift:3 pinter:13 york:1 latency:1 neocortex:1 encyclopedia:1 processed:1 category:1 simplest:1 inhibitory:4 per:4 discrete:2 darling:7 vol:2 srinivasan:1 lotka:2 luminance:3 relaxation:1 angle:1 micron:1 oscillation:1 kji:1 opto:1 bound:1 layer:2 fan:3 activity:3 preperation:1 strength:4 optic:1 constraint:1 speed:1 nmos:3 extremely:1 optical:1 department:1 membrane:2 bethesda:1 cable:1 biologically:1 depletion:4 computationally:1 equation:4 ln:1 previously:1 mechanism:1 end:1 gaussians:1 gate:3 especially:1 society:1 psychophysical:1 capacitance:2 pergamon:1 realized:1 receptive:3 dependence:5 md:1 exhibit:1 elicits:1 lateral:3 capacity:1 shifter:4 assuming:1 code:1 polarity:4 ratio:1 robert:2 negative:1 implementation:21 design:7 conversion:1 upper:2 vertical:1 neuron:1 neurobiology:1 arbitrary:1 intensity:4 introduced:1 required:2 connection:5 coherent:1 pattern:5 including:1 memory:6 suitable:1 interconnects:1 technology:3 lightness:1 axis:1 excitable:1 iength:1 multiplication:1 relative:1 law:1 fully:1 interesting:1 proportional:2 digital:2 sufficient:1 lo:1 excitatory:5 ellias:2 compatible:1 arriving:1 wide:2 neighbor:1 absolute:1 feedback:4 depth:1 contour:1 sensory:4 forward:1 made:1 adaptive:2 transaction:1 compact:1 neurobiological:1 logic:2 global:3 handbook:1 xi:11 iterative:1 channel:1 transfer:1 contributes:1 noise:1 carpenter:2 site:1 fig:3 exponential:1 xl:1 kandel:1 rail:5 theorem:1 specific:1 inset:3 showing:2 reverberating:1 decay:1 admits:1 physiological:1 evidence:1 suited:2 visual:7 saturating:1 bioi:2 sized:1 man:1 content:3 feasible:1 except:1 conductor:1 total:1 parasitic:1 preparation:1 tested:1 |
7,232 | 990 | Comparing the prediction accuracy of
artificial neural networks and other
statistical models for breast cancer
survival
Harry B. Burke
Department of Medicine
New York Medical College
Valhalla, NY 10595
David B. Rosen
Department of Medicine
New York Medical College
Valhalla, NY 10595
Philip H. Goodman
Department of Medicine
University of Nevada School of Medicine
Reno, Nevada 89520
Abstract
The TNM staging system has been used since the early 1960's
to predict breast cancer patient outcome. In an attempt to increase prognostic accuracy, many putative prognostic factors have
been identified. Because the TNM stage model can not accommodate these new factors, the proliferation of factors in breast
cancer has lead to clinical confusion. What is required is a new
computerized prognostic system that can test putative prognostic
factors and integrate the predictive factors with the TNM variables in order to increase prognostic accuracy. Using the area under the curve of the receiver operating characteristic, we compare
the accuracy of the following predictive models in terms of five
year breast cancer-specific survival: pTNM staging system, principal component analysis, classification and regression trees, logistic
regression, cascade correlation neural network, conjugate gradient
descent neural, probabilistic neural network, and backpropagation
neural network. Several statistical models are significantly more ac-
1064
Harry B. Burke, David B. Rosen, Philip H. Goodman
curate than the TNM staging system. Logistic regression and the
backpropagation neural network are the most accurate prediction
models for predicting five year breast cancer-specific survival
1
INTRODUCTION
For over thirty years measuring cancer outcome has been based on the TNM staging
system (tumor size, number of lymph nodes with metastatic disease, and distant
metastases) (Beahr et. al., 1992). There are several problems with this model
(Burke and Henson , 1993). First, it is not very accurate, for breast cancer it is
44% accurate. Second its accuracy can not be improved because predictive variables can not be added to the model. Third, it does not apply to all cancers. In
this paper we compare computerized prediction models to determine if they can
improve prognostic accuracy. Artificial neural networks (ANN) are a class of nonlinear regression and discrimination models. ANNs are being used in many areas
of medicine, with several hundred articles published in the last year. Representative areas of research include anesthesiology (Westenskow et. al., 1992) , radiology
(Tourassi et. al. , 1992) , cardiology (Leong and Jabri, 1982) , psychiatry (Palombo,
1992), and neurology (Gabor and Seyal, 1992). ANNs are being used in cancer
research including image processing (Goldberg et. al., 1992) , analysis of laboratory data for breast cancer diagnosis (0 Leary et. al., 1992), and the discovery of
chemotherapeutic agents (Weinstein et . al., 1992). It should be pointed out that
the analyses in this paper rely upon previously collected prognostic factors. These
factors were selected for collection because they were significant in a generalized
linear model such as the linear or logistic models . There is no predictive model that
can improve upon linear or logistic prediction models when the predictor variables
meet the assumptions of these models and there are no interactions. Therefore
he objective of this paper is not to outperform linear or logistic models on these
data. Rather, our objective is to show that, with variables selected by generalized
linear models, artificial neural networks can perform as well as the best traditional
models . There is no a priori reason to believe that future prognostic factors will
be binary or linear, and that there will not be complex interactions between prognostic factors. A further objective of this paper is to demonstrate that artificial
neural networks are likely to outperform the conventional models when there are
unanticipated nonmonotonic factors or complex interactions.
2
2.1
METHODS
DATA
The Patient Care Evaluation (PCE) data set is collected by the Commission on
Cancer of the American College of Surgeons (ACS). The ACS, in October 1992,
requested cancer information from hospital tumor registries in the United States.
The ACS asked for the first 25 cases of breast cancer seen at that institution in 1983,
and it asked for follow up information on each of these 25 patients through the date
of the request. These are only cases of first breast cancer. Follow-up information
included known deaths. The PCE data set contains, at best, eight year follow-up.
Prediction Accuracy of Models for Breast Cancer Survival
1065
We chose to use a five year survival end-point. This analysis is for death due to
breast cancer, not all cause mortality.
For this analysis cases with missing data, and cases censored before five years, are
not included so that the prediction models can be compared without putting any
prediction model at a disadvantage. We randomly divided the data set into training,
hold-out, and testing subsets of 3,100, 2,069, and 3,102 cases, respectively.
2.2
MODELS
The TMN stage model used in this analysis is the pathologic model (pTNM) based
on the 1992 American Joint Committee on Cancer's Manual for the Staging of
Cancer (Beahr et. al., 1992). The pathologic model relies upon pathologically determined tumor size and lymph nodes, this contrasts with clinical staging which
relies upon the clinical examination to provide tumor size and lymph node information. To determine the overall accuracy of the TNM stage model we compared
the model's prediction for each patient, where the individual patient's prediction
is the fraction of all the patients in that stage who survive, to each patient's true
outcome.
Principal components analysis, is a data reduction technique based on the linear
combinations of predictor variables that minimizes the variance across patients (Jollie, 1982). The logistic regression analysis is performed in a stepwise manner, without interaction terms, using the statistical language S-PLUS (S-PLUS, 1992), with
the continuous variable age modeled with a restricted cubic spline to avoid assuming
linearity (Harrell et. al., 1988). Two types of Classification and Regression Tree
(CART) (Breiman et. al., 1984) analyses are performed using S-PLUS. The first
was a 9-node pruned tree (with 10-fold cross validation on the deviance), and the
second was a shrunk tree with 13.7 effective nodes.
The multilayer perceptron neural network training in this paper is based on the
maximum likelihood function unless otherwise stated, and backpropagation refers
to gradient descent. Two neural networks that are not multilayer perceptrons are
tested. They are the Fuzzy ARTMAP neural network (Carpenter et. al., 1991) and
the probabilistic neural network (Specht, 1990).
2.3
ACCURACY
The measure of comparative accuracy is the area under the curve of the receiver
operating characteristic (Az) . Generally, the Az is a nonparametric measure of
discrimination. Square error summarizes how close each patient's predicted value is
to its true outcome. The Az measures the relative goodness of the set of predictions
as a whole by comparing the predicted probability of each patient with that of all
other patients. The computational approach to the Az that employs the trapezoidal
approximation to the area under the receiver operating characteristic curve for
binary outcomes was first reported by Bamber (Bamber, 1975), and later in the
medical literature by Hanley (Hanley and McNeil, 1982). This was extended by
Harrell (Harrell et. al., 1988) to continuous outcomes.
Harry B. Burke, David B. Rosen, Philip H. Goodman
1066
Table 1: PCE 1983 Breast Cancer Data: 5 Year Survival Prediction, 54 Variables.
PREDICTION MODEL
pTNM Stages
Principal Components Analysis
CART, pruned
CART, shrunk
Stepwise Logistic regression
Fuzzy ARTMAP ANN
Cascade correlation ANN
Conjugate gradient descent ANN
Probabilistic ANN
Backpropagation ANN
ACCURACY?
SPECIFICATIONS
.720
.714
.753
.762
.776
.738
.761
.774
.777
.784
O,I,I1A,I1B,IIIA,I1IB,IV
one scaling iteration
9 nodes
13.7 nodes
with cubic splines
54-F2a, 128-1
54-21-1
54-30-1
bandwidth = 16s
54-5-1
* The area under the curve of the receiver operating characteristic.
3
RESULTS
All results are based on the independent variable sample not used for training (i.e.,
the testing data set), and all analyses employ the same testing data set. Using
the PCE breast cancer data set, we can assess the accuracy of several prediction
models using the most powerful of the predictor variables available in the data set
(See Table 1).
Principal components analysis is not expected to be a very accurate model; with
one scaling iteration, its accuracy is .714. Two types of classification and regression trees (CART), pruned and shrunk, demonstrate accuracies of .753 and .762,
respectively. Logistic regression with cubic splines for age has an accuracy of .776.
In addition to the backpropagation neural network and the probabilistic neural network, three types of neural networks are tested. Fuzzy ARTMAP's accuracy is
the poorest at .738. It was too computationally intensive to be a practical model.
Cascade-correlation and conjugate gradient descent have the potential to do as well
as backpropagation. The PNN accuracy is .777. The PNN has many interesting
features, but it also has several drawbacks including its storage requirements. The
backpropagation neural network's accuracy is .784.4.
4
DISCUSSION
For predicting five year breast cancer-specific survival, several computerized prediction models are more accurate than the TNM stage system, and artificial neural
networks are as good as the best traditional statistical models.
References
Bamber D (1975). The area above the ordinal dominance graph and the area below
the receiver operating characteristic. J Math Psych 12:387-415.
Beahrs OH, Henson DE, Hutter RVP, Kennedy BJ (1992). Manual for staging of
Prediction Accuracy of Models for Breast Cancer Survival
1067
cancer, 4th ed. Philadelphia: JB Lippincott.
Burke HB, Henson DE (1993) . Criteria for prognostic factors and for an enhanced
prognostic system. Cancer 72:3131-5.
Breiman L, Friedman JH, Olshen RA (1984). Classification and Regression Trees.
Pacific Grove, CA: Wadsworth and Brooks/Cole.
Carpenter GA, Grossberg S, Rosen DB (1991). Fuzzy ART: Fast stable learning
and categorization of analog patterns by an adaptive resonance system. Neural
Networks 4:759-77l.
Gabor AJ, M. Seyal M (1992) . Automated interictal EEG spike detection using
artificial neural networks. Electroencephalogr Clin Neurophysiology 83 :271-80.
Goldberg V, Manduca A, Ewert DL (1992). Improvement in specificity of ultrasonography for diagnosis of breast tumors by means of artificial intelligence. Med
Phys 19:1275-8l.
Hanley J A, McNeil BJ (1982). The meaning of the use of the area under the receiver
operating characteristic (ROC) curve. Radiology 143:29-36.
Harrell FE, Lee KL, Pollock BG (1988). Regression models in clinical studies:
determining relationships between predictors and response. J Natl Cancer Instit
80:1198-1202.
Jollife IT (1986). Principal Component Analysis. New York: Springer-Verlag, 1986.
Leong PH, J abri MA (1982). MATIC - an intracardiac tachycardia classification
system. PACE 15:1317-31,1982.
O'Leary TJ, Mikel UV, Becker RL (1992). Computer-assisted image interpretation:
use of a neural network to differentiate tubular carcinoma from sclerosing adenosis.
Modern Pathol 5:402-5.
Palombo SR (1992). Connectivity and condensation in dreaming. JAm Psychoanal
Assoc 40:1139-59.
S-PLUS (1991), v 3.0. Seattle, WA; Statistical Sciences, Inc.
Specht DF (1990). Probabilistic neural networks. Neural Networks 3:109-18.
Tourassi GD, Floyd CE, Sostman HD, Coleman RE (1993). Acute pulmonary
embolism: artificial neural network approach for diagnosis. Radiology 189:555-58.
Weinstein IN, Kohn KW, Grever MR et. al. (1992) Neural computing in cancer
drug development: predicting mechanism of action. Science 258:447-51.
Westenskow DR, Orr JA, Simon FH (1992) . Intelligent alarms reduce anesthesiologist's response time to critical faults. Anesthesiology 77:1074-9, 1992.
Learning with Product Units
Laurens R. Leerink
Australian Gilt Securities LTD
37-49 Pitt Street
NSW 2000, Australia
laurens@sedal.su.oz.au
Bill G. Horne
NEC Research Institute
4 Independence Way
Princeton, NJ 08540, USA
horne@research.nj.nec.com
C. Lee Giles
NEC Research Institute
4 Independence Way
Princeton, NJ 08540, USA
giles@research.nj.nec.com
Marwan A. Jabri
Department of Electrical Engineering
The University of Sydney
NSW 2006, Australia
marwan@sedal.su.oz.au
Abstract
Product units provide a method of automatically learning the
higher-order input combinations required for efficient learning in
neural networks. However, we show that problems are encountered when using backpropagation to train networks containing
these units. This paper examines these problems, and proposes
some atypical heuristics to improve learning. Using these heuristics
a constructive method is introduced which solves well-researched
problems with significantly less neurons than previously reported.
Secondly, product units are implemented as candidate units in the
Cascade Correlation (Fahlman & Lebiere, 1990) system. This resulted in smaller networks which trained faster than when using
sigmoidal or Gaussian units.
1
Introduction
It is well-known that supplementing the inputs to a neural network with higher-order
combinations ofthe inputs both increases the capacity of the network (Cover, 1965)
and the the ability to learn geometrically invariant properties (Giles & Maxwell,
538
Laurens Leerink, C. Lee Giles, Bill G. Home, Marwan A. Jabri
1987). However, there is a combinatorial explosion of higher order terms as the
number of inputs to the network increases. Yet in order to implement a certain
logical function, in most cases only a few of these higher order terms are required
(Redding et al., 1993).
The product units (PUs) introduced by (Durbin & Rumelhart, 1989) attempt to
make use of this fact. These networks have the advantage that, given an appropriate
training algorithm, the units can automatically learn the higher order terms that
are required to implement a specific logical function.
In these networks the hidden layer units compute the weighted product ofthe inputs,
that is
N
N
II X~i
instead of
i=l
2:XiWi
(1)
i=l
as in standard networks. An additional advantage of PUs is the increased information capacity of these units compared to standard summation networks. It is
approximately 3N (Durbin & Rumelhart, 1989), compared to 2N for a single
threshold logic function (Cover, 1965), where N is the number of inputs to the
unit.
The larger capacity means that the same functions can be implemented by networks
containing less units. This is important for certain applications such as speech
recognition where the data bandwidth is high or if realtime implementations are
desired.
When PUs are used to process Boolean inputs, best performance is obtained
(Durbin & Rumelhart, 1989) by using inputs of {+1, -I}. If the imaginary component is ignored, with these inputs, the activation function is equivalent to a cosine
summation function with {-1,+1} inputs mapped {I,D} (Durbin & Rumelhart,
1989). In the remainder of this paper the terms product unit (PU) and cos{ine)
unit will be used interchangeably as all the problems examined have Boolean inputs.
2
Learning with Product Units
As the basic mechanism of a PU is multiplicative instead of additive, one would
expect that standard neural network training methods and procedures cannot be
directly applied when training these networks. This is indeed the case. If a neural
network simulation environment is available the basic functionality of a PU can be
obtained by simply adding the cos function cos( 1(" * input) to the existing list of
transfer functions. This assumes that Boolean mappings are being implemented
and the appropriate {-1,+1} - {I,D} mapping has been performed on the input
vectors. However, if we then attempt to train a network on on the parity-6 problem
shown in (Durbin & Rumelhart, 1989), it is found that the standard backpropagat ion (BP) algorithm simply does not work. We have found two main reasons for
this.
The first is weight initialization. A typical first step in the backpropagation procedure is to initialize all weights to small random values. The main reason for this
is to use the dynamic range of the sigmoid function and it's derivative. However,
the dynamic range of a PU is unlimited. Initializing the weights to small random
539
Learning with Product Units
values results in an input to the unit where the derivative is small. So apart from
choosing small weights centered around ml" with n = ?1, ?2, ... this is the worst
possible choice. In our simulations weights were initialized randomly in the range
[-2,2]. In fact, learning seems insensitive to the size of the weights, as long as they
are large enough.
The second problem is local minima. Previous reports have mentioned this problem, (Lapedes & Farber, 1987) commented that "using sin's often leads to numerical problems, and nonglobal minima, whereas sigmoids seemed to avoid such
problems". This comment summarizes our experience of training with PUs. For
small problems (less than 3 inputs) backpropagation provides satisfactory training.
However, when the number of inputs are increased beyond this number, even with
th: .weight initialization in the correct range, training usually ends up in a local
mInIma.
3
Training Algorithms
With these aspects in mind, the following training algorithms were evaluated: online
and batch versions of Backpropagation (BP), Simulated Annealing (SA), a Random
Search Algorithm (RSA) and combinations of these algorithms.
BP was used as a benchmark and for use in combination with the other algorithms.
The Delta-Bar-Delta learning rate adaptation rule (Jacobs, 1988) was used along
with the batch version of BP to accelerate convergence, with the parameters were
set to 0
0.35, K,
0.05 and ?
0.90. RSA is a global search method (i.e.
the whole weight space is explored during training). Weights are randomly chosen
from a predefined distribution, and replaced if this results in an error decrease. SA
(Kirkpatrick et aI., 1983) is a standard optimization method. The operation of SA
is similar to RSA, with the difference that with a decreasing probability solutions
are accepted which increase the training error. The combination of algorithms were
chosen (BP & SA, BP & RSA) to combine the benefits of global and local search.
Used in this manner, BP is used to find the local minima. If the training error at
the minima is sufficiently low, training is terminated. Otherwise, the global method
initializes the weights to another position in weight space from which local training
can continue.
=
=
=
The BP-RSA combination requires further explanation. Several BP-(R)SA combinations were evaluated, but best performance was obtained using a fixed number of
iterations of BP (in this case 120) along with one initial iteration of RSA. In this
manner BP is used to move to the local minima, and if the training error is still
above the desired level the RSA algorithm generates a new set of random weights
from which BP can start again.
The algorithms were evaluated on two problems, the parity problem and learning all
logical functions of 2 and 3 inputs. The infamous parity problem is (for the product
unit at least) an appropriate task. As illustrated by (Durbin & Rumelhart, 1989),
this problem can be solved by one product unit. The question is whether the
training algorithms can find a solution. The target values are {-1, +1}, and the
output is taken to be correct if it has the correct sign. The simulation results are
shown in Table 1. It should be noted that one epoch of both SA and RSA involves
540
Laurens Leerink, C. Lee Giles, Bill G. Home, Marwan A. Jahri
relaxing the network across the training set for every weight, so in computational
terms their nepoeh values should be multiplied by a factor of (N + 1).
Parity
N
6
8
10
Online BP
Batch BP
neonv
nepoeh
n eonv
10
8
6
30.4
101.3
203.3
7
2
0
SA
RSA
nepoeh
neonv
nepoeh
neon v
nepoeh
34
700
10
10
10
12.6
52.8
99.9
10
10
10
15.2
45.4
74.1
-
Table 1: The parity N problem: The table shows neon v the number of runs out of
10 that have converged and nepoeh' the average number of training epochs required
when training converged.
For the parity problem it is clear that local learning alone does not provide good
convergence. For this problem, global search algorithms have the following advantages: (1) The search space is bounded (all weights are restricted to [-2, +2]) (2)
The dimension of search space is low (maximum of 11 weights for the problems
examined). (3) The fraction of the weight space which satisfies the parity problem
relative to the total bounded weight space is high.
In a second set of simulations, one product unit was trained to calculate all 2(2 N )
logical functions of the N input variables. Unfortunately, this is only practical for
N E {2,3} . For N = 2 there are only 16 functions, and a product unit has no
problem learning all these functions rapidly with all four training algorithms. In
comparison a single summation unit can learn 14 (not the XOR & XNOR functions).
For N =3, a product unit is able to implement 208 of the 256 functions, while a single
summation unit could only implement 104. The simulation results are displayed in
Table 2.
Online BP
Batch BP
BP-RSA
Table 2: Learning all logical functions of 3 inputs: The rows display nlogie , the
average number of logical functions implemented by a product unit and nepoeh, the
number of epochs required for convergence. Ten simulations were performed for
each of the 256 logical functions , each for a maximum of 1,000 iterations.
4
Constructive Learning with Product Units
Selecting the optimal network architecture for a specific application is a nontrivial
and time-consuming task, and several algorithms have been proposed to automate
this process. These include pruning methods and growing algorithms. In this section
a simple method is proposed for adding PUs to the hidden layer of a three layer
network. The output layer contains a single sigmoidal unit .
Several constructive algorithms proceed by freezing a subset of the weights and
limiting training to the newly added units. As mentioned earlier, for PUs a global
541
Learning with Product Units
300
Tiling AI orithm
Upstart AI orithm
81M using Pr
Units
~
I-t--<
>S-t
250
i!:
0
~c
200
.!;
<II
C
e::l
150
CI)
c
'15
~
D
E
::l
Z
100
50
200
400
600
800
Number of patterns (2"N)
1000
1200
Figure 1: The number of units required for learning the random mapping problems
by the 'Tiling', 'Upstart' and SIM algorithms.
search is required to solve the local-minima problems. Freezing a subset of the
weights restricts the new solution to an affine subset of the existing weight space,
often resulting in non-minimal networks (Ash, 1989). For this reason a simple
incremental method (SIM) was implemented which retains the global search for all
weights during the whole training process. The method used in our simulations is
as follows:
? Train a network using the BP-RSA combination on a network with a specified minimum number of hidden PUs.
? If there is no convergence within a specified number of epochs, add a PU to
the network. Reinitialize weights and continue training with the BP-RSA
combination .
? Repeat process until a solution is found or the network has grown a predetermined maximum size.
The method of (Ash , 1989) was also evaluated, where neurons with small weights
were added to a network according to certain criteria. The SIM performed better,
possibly because of the global search performed by the RSA step.
The 'Upstart' (Frean, 1990) and 'Tiling' (Mezard & Nadal, 1989) constructive
algorithms were chosen as benchmarks. A constructive PU network was trained
on two problems described in these papers, namely the parity problem and the
random mapping problem. In (Frean, 1990) it was reported that the Upstart
542
Laurens Leerink, C. Lee Giles, Bill G. Home, Marwan A. Jabri
algorithm required N units for all parity N problems, and 1,000 training epochs
were sufficient for all values of N except N = 10, which required 10,000. As seen
earlier, one PU is able to perform any parity function, and SIM required an an
average of 74.1 iterations for N = 6,8,10.
The random mapping problem is defined by assigning each of the 2N patterns its
target { -1, +I} with 50% probability. This is a difficult problem, due to the absence
of correlations and structure in the input. As in (Frean, 1990; Mezard & Nadal,
1989) the average of 25 runs were performed, each on a different training set. The
number of units required by SIM is plotted in Figure 1. The values for the Tiling
and Upstart algorithms are approximate and were obtained through inspection from
a similar graph in (Frean, 1990).
5
U sing Cosine Candidate Units in Cascade Correlation
Initially we wanted to compare the performance of SIM with the well-known
'cascade-correlation' (CC) algorithm of (Fahlman & Lebiere, 1990). However, the
network architectures differ and a direct comparison between the number of units in
the respective architectures does not reflect the efficiency of the algorithms. Instead,
it was decided to integrate PUs into the CC system as candidate units.
For these simulations a public domain version of CC was used (White, 1993) which
supports four different candidate types; the asymmetric sigmoid, symmetric sigmoid, variable sigmoid and gaussian units. Facilities exist for either constructing
homogeneous networks by selecting one unit type, or training with a pool of different units allowing the construction of hybrid networks. It was thus relatively simple
to add PU candidate units to the system. Table 3 displays the results when CC was
trained on the random logic problem using three types of homogeneous candidate
units.
N
7
8
9
10
CC Sigmoid
CC Gauss
CC PU
nunih
nepOCh6
nunih
nepoch,
nunih
n epoch6
6.6
12.1
20.5
32.9
924.5
1630.9
2738.3
4410.9
6.7
11.5
18.4
30.2
642.6
1128.2
1831.1
2967.6
5.7
9.9
16.4
26.6
493.8
833.8
1481.8
2590.8
Table 3: Learning random logic functions of N inputs: The table shows nunih'
the average number of units required and nepoch6' the average number of training
epochs required for convergence of CC using sigmoidal, Gaussian and PU candidate
units . Figures are based on 25 simulations.
In a separate experiment the performance of hybrid networks were re-evaluated on
the same random logic problem. To enable a fair competition between candidate
units of different types, the simulations were run with 40 candidate units, 8 of each
type. The simulations were evaluated on 25 trails for each of the random mapping
problems (7,8,9 and 10 inputs, a total of 1920 input vectors). In total 1460 hidden
units were allocated, and in all cases PU candidate units were chosen above units
of the 4 other types during the competitive stage. During this comparison all
Learning with Product Units
543
parameters were set to default values, i.e. the weights of the PU candidate units
were random numbers initialized in the range of [-1, +1]. As discussed earlier, this
puts the PUs at a slight disadvantage as their optimum range is [-2, +2].
6
Discussion
The BP-RSA combination is in effect equivalent to the 'local optimization with
random restarts' process discussed by (Karmarkar & Karp, 1982), where the local
optimization is this case is performed by the BP algorithm. They reported that
for certain problems where the error surface was 'exceedingly mountainous', multiple random-start local optimization outperformed more sophisticated methods.
We hypothesize that adding PUs to a network makes the error surface sufficiently
mountainous so that a global search is required.
As expected, the higher separating capacity of the PU enables the construction of
networks with less neurons than those produced by the Tiling and Upstart algorithms. The fact that SIM works this well is mainly a result of the error surface; the
surface is so irregular that even training a network of fixed architecture is best done
by reinitializing the weights if convergence does not occur within certain bounds.
This again is in accordance with the results of (Karmarkar & Karp, 1982) discussed
above .
When used in CC we hypothesize that there are three main reasons for the choice
of PUs above any of the other types during the competitive learning phase. Firstly,
the higher capacity (in a information capacity sense) of the PUs allows a better
correlation with the error signal. Secondly, having N competing candidate units is
equivalent to selecting the best of N random restarts, and performs the required
global search. Thirdly, although the error surface of networks with PUs contains
more local minima than when using standard transfer functions, the surface is locally
smooth. This allows effective use of higher-order error derivatives, resulting in fast
convergence by the quickprop algorithm.
In (Dawson & Schopflocher, 1992) it was shown that networks with Gaussian units
train faster and require less units than networks with standard sigmoidal units.
This is supported by our results shown in Table 3. However, for the problem
examined, PUs outperform Gaussian units by approximately the same margin as
Gaussian units outperform sigmoidal units. It should also be noted that these
problems where not chosen for their suitability for PUs. In fact, if the problems are
symmetric/regular the difference in performance is expected to increase.
7
Conclusion
Of the learning algorithms examined BP provides the fastest training, but is prone
to nonglobal minima. On the other hand, global search methods are impractical
for larger networks. For the problems examined, a combination of local and global
search methods were found to perform best. Given a network containing PUs, there
are some atypical heuristics that can be used: (a) correct weight initialization (b)
reinitialization of the weights if convergence is not rapidly reached. In addition,
the representational power of PUs have enabled us to solve standard problems
544
lAurens Leerink, C. Lee Giles, Bill G. Home, Marwan A. labri
using significantly smaller networks than previously reported, using a very simple
constructive method. When implemented in the CC architecture, for the problems
examined PUs resulted in smaller networks which trained faster than other units.
When included in a pool of competing candidate units, simulations showed that in
all cases PU candidate units were preferred over candidate units of the other four
types.
References
Ash, T. (1989) . Dynamic node creation in backpropagation networks. Connection
Science, 1 (4), 365-375.
Cover, T. (1965) . Geometrical and statistical properties of systems of linear inequalities with applications in pattern recognition . IEEE Transactions on Electronic
Computers, 14, 326-334.
Dawson, M. & Schopflocher, D. (1992) . Modifying the generalized delta rule to train
networks of nonmonotonic processors for pattern classification. Connection
Science, 4, 19-31.
Durbin, R. & Rumelhart, D. (1989). Product units: A computationally powerful and biologically plausible extension to backpropagation networks. Neural
Computation, 1, 133- 142.
Fahlman, S. & Lebiere, C . (1990) . The cascade-correlation learning architecture.
In Touretzky, D. (Ed.), Advances in Neural Information Processing Systems,
volume 2, (pp. 524-532)., San Mateo. (Denver 1989), Morgan Kaufmann.
Frean, M. (1990). The upstart algorithm: A method for constructing and training
feedforward neural networks. Neural Computation, 2, 198-209.
Giles, C. & Maxwell, T . (1987). Learning, invariance, and generalization in highorder neural networks. Applied Optics, 26(23),4972-4978 .
Jacobs, R. (1988). Increased rates of convergence through learning rate adaptation .
Neural Networks, 1, 295-307.
Karmarkar , N. & Karp, R . (1982). The differencing method of set partitioning.
Technical Report UCB/CSD 82/113, Computer Science Division, University of
California, Berkeley, California.
Kirkpatrick, S., Jr., C. G ., , & Vecchi, M. (1983) . Optimization by simulated
annealing. Science, 220. Reprinted in (?) .
Lapedes, A. & Farber, R. (1987) . Nonlinear signal processing using neural networks: Prediction and system modelling. Technical Report LA-UR-87-2662,
Los Alamos National Laboratory, Los Alamos, NM.
Mezard, M. & Nadal, J.-P. (1989). Learning in feedforward layered networks: The
tiling algorithm. Journal of Physics A, 22, 2191-2204.
Redding, N., Kowalczyk, A., & Downs, T . (1993). A constructive higher order
network algorithm that is polynomial-time. Neural Networks, 6,997.
White, M. (1993). A public domain C implement ion of the Cascade Correlation algorithm. Department of Computer Science, Carnegie Mellon University, Pittsburgh , PA.
| 990 |@word neurophysiology:1 version:3 polynomial:1 seems:1 prognostic:11 simulation:12 quickprop:1 jacob:2 nsw:2 accommodate:1 reduction:1 initial:1 contains:3 united:1 selecting:3 lapedes:2 imaginary:1 existing:2 comparing:2 com:2 activation:1 yet:1 assigning:1 distant:1 additive:1 numerical:1 predetermined:1 enables:1 wanted:1 hypothesize:2 discrimination:2 alone:1 intelligence:1 selected:2 inspection:1 coleman:1 institution:1 provides:2 math:1 node:8 sigmoidal:5 seyal:2 firstly:1 five:5 along:2 direct:1 weinstein:2 combine:1 manner:3 ra:1 indeed:1 expected:3 proliferation:1 growing:1 decreasing:1 automatically:2 researched:1 supplementing:1 horne:2 linearity:1 bounded:2 what:1 xiwi:1 minimizes:1 psych:1 fuzzy:4 nadal:3 nj:4 impractical:1 berkeley:1 every:1 assoc:1 partitioning:1 unit:62 medical:3 mountainous:2 before:1 engineering:1 local:13 accordance:1 infamous:1 meet:1 approximately:2 chose:1 plus:4 au:2 initialization:3 examined:6 mateo:1 relaxing:1 co:3 fastest:1 range:6 grossberg:1 practical:2 thirty:1 decided:1 testing:3 rsa:14 implement:5 backpropagation:13 procedure:2 area:9 drug:1 cascade:8 significantly:3 gabor:2 deviance:1 refers:1 specificity:1 regular:1 cardiology:1 tmn:1 cannot:1 close:1 ga:1 layered:1 storage:1 put:1 conventional:1 bill:5 equivalent:3 missing:1 artmap:3 orithm:2 examines:1 rule:2 oh:1 hd:1 enabled:1 bamber:3 limiting:1 enhanced:1 target:2 construction:2 homogeneous:2 goldberg:2 trail:1 pa:1 rumelhart:7 recognition:2 trapezoidal:1 asymmetric:1 electrical:1 initializing:1 worst:1 solved:1 calculate:1 decrease:1 disease:1 mentioned:2 environment:1 asked:2 dynamic:3 highorder:1 trained:5 f2a:1 lippincott:1 predictive:4 surgeon:1 upon:4 anesthesiologist:1 efficiency:1 creation:1 division:1 accelerate:1 joint:1 grown:1 train:5 dreaming:1 fast:2 effective:2 artificial:8 labri:1 outcome:6 nonmonotonic:2 choosing:1 heuristic:3 larger:2 metastatic:1 solve:2 plausible:1 otherwise:2 ability:1 radiology:3 online:3 differentiate:1 advantage:3 nevada:2 interaction:4 product:18 remainder:1 adaptation:2 date:1 rapidly:2 representational:1 oz:2 ine:1 competition:1 az:4 los:2 seattle:1 convergence:9 requirement:1 optimum:1 comparative:1 categorization:1 incremental:1 ac:4 frean:5 school:1 sa:7 sim:7 sydney:1 pathologic:2 predicted:2 tourassi:2 implemented:6 australian:1 involves:1 differ:1 solves:1 laurens:6 drawback:1 farber:2 functionality:1 correct:4 shrunk:3 modifying:1 centered:1 australia:2 enable:1 public:2 jam:1 ja:1 require:1 generalization:1 suitability:1 secondly:2 summation:4 extension:1 assisted:1 burke:5 hold:1 around:1 sufficiently:2 mapping:6 predict:1 bj:2 pitt:1 automate:1 early:1 valhalla:2 fh:1 outperformed:1 combinatorial:1 cole:1 electroencephalogr:1 weighted:1 gaussian:6 rather:1 avoid:2 breiman:2 karp:3 improvement:1 modelling:1 likelihood:1 mainly:1 contrast:1 psychiatry:1 sense:1 initially:1 hidden:4 overall:1 classification:6 priori:1 development:1 resonance:1 art:1 proposes:1 initialize:1 wadsworth:1 having:1 abri:1 kw:1 survive:1 future:1 rosen:4 jb:1 metastasis:1 spline:3 intelligent:1 employ:2 report:3 modern:1 randomly:3 few:1 resulted:2 national:1 individual:1 replaced:1 phase:1 attempt:3 friedman:1 detection:1 evaluation:1 kirkpatrick:2 natl:1 tj:1 staging:7 predefined:1 accurate:5 grove:1 explosion:1 censored:1 experience:1 respective:1 unless:1 tree:6 iv:1 initialized:2 re:2 desired:2 plotted:1 minimal:1 hutter:1 increased:3 earlier:3 giles:8 boolean:3 cover:3 disadvantage:2 measuring:1 goodness:1 retains:1 subset:4 hundred:1 predictor:4 alamo:2 too:1 commission:1 reported:5 gd:1 schopflocher:2 probabilistic:5 lee:6 physic:1 pool:2 leary:2 connectivity:1 again:2 mortality:1 reflect:1 containing:3 nm:1 possibly:1 dr:1 american:2 derivative:3 potential:1 de:2 orr:1 harry:3 inc:1 ewert:1 bg:1 performed:8 later:1 multiplicative:1 i1a:1 start:2 curate:1 competitive:2 reached:1 simon:1 ass:1 square:1 accuracy:19 xor:1 variance:1 characteristic:6 who:1 kaufmann:1 ofthe:2 redding:2 produced:1 computerized:3 kennedy:1 cc:10 published:1 processor:1 converged:2 anns:2 phys:1 touretzky:1 manual:2 ed:2 pp:1 lebiere:3 newly:1 logical:7 sophisticated:1 maxwell:2 higher:9 follow:3 restarts:2 response:2 improved:1 evaluated:6 done:1 stage:7 correlation:10 until:1 hand:1 freezing:2 su:2 nonlinear:2 logistic:8 aj:1 believe:1 usa:2 effect:1 true:2 facility:1 symmetric:2 laboratory:2 death:2 satisfactory:1 illustrated:1 xnor:1 white:2 floyd:1 interchangeably:1 sin:1 during:5 noted:2 cosine:2 criterion:2 generalized:3 demonstrate:2 confusion:1 performs:1 geometrical:1 image:2 meaning:1 sigmoid:5 rl:1 denver:1 insensitive:1 volume:1 thirdly:1 analog:1 he:1 interpretation:1 discussed:3 slight:1 sedal:2 significant:1 mellon:1 ai:3 uv:1 pointed:1 language:1 henson:3 specification:1 stable:1 operating:6 acute:1 surface:6 pu:32 add:2 showed:1 apart:1 verlag:1 certain:5 inequality:1 binary:2 continue:2 dawson:2 fault:1 seen:2 minimum:10 additional:1 care:1 morgan:1 mr:1 pathol:1 determine:2 signal:2 ii:2 multiple:1 smooth:1 technical:2 faster:3 clinical:4 cross:1 tubular:1 long:1 divided:1 prediction:16 regression:11 basic:2 breast:16 patient:11 multilayer:2 intracardiac:1 df:1 iteration:6 pnn:2 ion:2 irregular:1 addition:2 condensation:1 whereas:1 annealing:2 allocated:1 goodman:3 sr:1 comment:1 cart:4 med:1 db:1 leong:2 feedforward:2 enough:1 hb:1 automated:1 independence:2 architecture:6 identified:1 bandwidth:2 registry:1 rvp:1 reduce:1 upstart:7 competing:2 reprinted:1 intensive:1 whether:1 kohn:1 becker:1 ltd:1 speech:1 york:3 cause:1 proceed:1 action:1 backpropagat:1 ignored:1 generally:1 clear:1 nonparametric:1 ten:1 westenskow:2 ph:1 locally:1 instit:1 outperform:4 exist:1 restricts:1 sign:1 delta:3 carcinoma:1 pace:1 diagnosis:3 carnegie:1 dominance:1 putting:1 commented:1 four:3 threshold:1 ce:1 ptnm:3 graph:2 mcneil:2 fraction:2 year:9 geometrically:1 run:3 powerful:2 electronic:1 putative:2 home:4 realtime:1 summarizes:2 scaling:2 poorest:1 bound:1 interictal:1 layer:4 display:2 fold:1 nonglobal:2 neon:2 encountered:1 durbin:7 nontrivial:1 occur:1 optic:1 bp:22 unlimited:1 generates:1 aspect:1 vecchi:1 pruned:3 relatively:1 department:5 pacific:1 according:1 request:1 combination:12 conjugate:3 jr:1 across:2 smaller:3 ur:1 biologically:1 restricted:2 invariant:1 pr:1 taken:1 computationally:2 previously:3 committee:1 manduca:1 mechanism:2 ordinal:1 mind:1 specht:2 end:2 tiling:6 available:2 operation:1 multiplied:1 apply:1 eight:1 appropriate:3 kowalczyk:1 batch:4 assumes:1 include:2 clin:1 medicine:5 hanley:3 objective:3 initializes:1 added:3 move:1 spike:1 question:1 reinitialize:1 traditional:2 gradient:4 pathologically:1 separate:1 ultrasonography:1 mapped:1 capacity:6 philip:3 street:1 simulated:2 separating:1 collected:2 reason:5 mikel:1 assuming:1 modeled:1 relationship:1 differencing:1 difficult:1 october:1 olshen:1 fe:1 unfortunately:1 stated:1 implementation:1 perform:3 allowing:1 neuron:3 benchmark:2 sing:1 descent:4 embolism:1 displayed:1 extended:1 unanticipated:1 princeton:2 david:3 introduced:2 namely:1 required:16 kl:1 specified:2 connection:2 security:1 lymph:3 california:2 brook:1 beyond:1 bar:1 able:2 below:1 pattern:5 usually:1 including:2 explanation:1 power:1 critical:1 rely:1 examination:1 predicting:3 hybrid:2 improve:3 anesthesiology:2 leerink:5 reno:1 philadelphia:1 epoch:6 literature:1 discovery:1 determining:1 relative:2 expect:1 interesting:1 age:2 validation:1 ash:3 integrate:2 agent:1 affine:1 sufficient:1 article:1 row:1 cancer:26 prone:1 repeat:1 last:1 fahlman:3 parity:10 supported:1 iiia:1 harrell:4 jh:1 perceptron:1 institute:2 benefit:1 curve:5 dimension:1 default:1 seemed:1 exceedingly:1 collection:1 adaptive:1 san:1 transaction:1 nunih:4 pruning:1 approximate:1 preferred:1 logic:4 ml:1 global:11 receiver:6 pittsburgh:1 marwan:6 consuming:1 neurology:1 matic:1 continuous:2 search:13 table:11 learn:3 transfer:2 ca:1 eeg:1 requested:1 complex:2 constructing:2 jabri:4 domain:2 tachycardia:1 main:3 csd:1 terminated:1 whole:3 alarm:1 fair:1 carpenter:2 representative:1 roc:1 cubic:3 ny:2 position:1 mezard:3 candidate:15 atypical:2 third:1 pollock:1 down:1 tnm:7 specific:5 list:1 explored:1 survival:8 dl:1 stepwise:2 beahr:2 adding:3 ci:1 nec:4 sigmoids:1 margin:1 simply:2 likely:1 palombo:2 springer:1 pulmonary:1 satisfies:1 relies:2 ma:1 ann:6 absence:1 included:3 determined:1 typical:1 except:1 principal:5 tumor:5 total:3 pce:4 hospital:1 accepted:1 gauss:1 invariance:1 la:1 ucb:1 perceptrons:1 college:3 support:1 constructive:7 karmarkar:3 tested:2 |
7,233 | 991 | Stochastic Dynamics of Three-State
Neural Networks
Toru Ohira
Sony Computer Science Laboratory
3-14-13 Higashi-gotanda,
Tokyo 141, Japan
ohira@csl.sony.co.jp
Jack D. Cowan
Depts. of Mathematics and Neurology
University of Chicago
Chicago, IL 60637
cowan@synapse.uchicago.edu
Abstract
We present here an analysis of the stochastic neurodynamics of
a neural network composed of three-state neurons described by
a master equation. An outer-product representation of the master equation is employed. In this representation, an extension of
the analysis from two to three-state neurons is easily performed.
We apply this formalism with approximation schemes to a simple three-state network and compare the results with Monte Carlo
simulations.
1
INTRODUCTION
Studies of single neurons or networks under the influence of noise have been a continuing item in neural network modelling. In particular, the analogy with spin
systems at finite temprature has produced many important results on networks of
two-state neurons. However, studies of networks of three-state neurons have been
rather limited (Meunier, Hansel and Verga, 1989). A master equation was introduced by Cowan (1991) to study stochastic neural networks. The equation uses the
formalism of "second quantization" for classical many-body systems (Doi, 1976a;
Grassberger and Scheunert, 1980), and was used to study networks of of two-state
neurons (Ohira and Cowan, 1993, 1994). In this paper, we reformulate the master
equation using an outer-product representation of operators and extend our previous analysis to networks of three-state neurons. A hierarchy of moment equations
for such networks is derived and approximation schemes are used to obtain equa-
272
Toru Ohira, Jack D. Cowan
Figure 1: Transition rates for a three-state neuron.
tions for the macroscopic activities of model networks. We compare the behavior
of the solutions of these equations with Monte Carlo simulations.
2
THE BASIC NEURAL MODEL
We first introduce the network described by the master equation. In this network
(Cowan, 1991), neurons at each site, say the ith site, are assumed to cycle through
three states: "quiescent", "activated" and "refractory", labelled 'qi', 'aj', and 'ri'
respectively. We consider four transitions: q --t a, l' --t a, a --t 1', and " --t q. Two of
these, q --t a and r --t a, are functions of the neural input current. We assume these
are smoothly increasing functions of the input current and denoted them by fh, and
(}2. The other two transition rates, a --t r, and r --t q, are defined as constants a and
/3. The resulting stochastic transition scheme is shown in Figure 1. We assume that
these transition rates depend only on the current state of the network and not on
past states, and that all neural state transitions are asynchronous. This Markovian
assumption is essential to the master equation description of this model.
We represent the state of each neuron by three-dimensional basis vectors using the
Dirac notation lai >, Iri > and Iqi >. They correspond, in more standard vector
notation, to:
Iq;
>=
(D,
la;>=
0),
I
h >= (
I
~)
o ?.
(1)
We define the inner product of these states as
< adai >=< qdqj >=< 1'jlri >= 1,
(2)
< qjlaj >=< ajlqi >=< rdai >=< ad"i >=< 1'dqi >=< Qil " j >= O.
(3)
Let the states (or configurations) of a network be represented by
product space of each neuron in the network.
{In>}, the direct
(4)
Let p[n, t] be the probability of finding the network in a particular state n at time
t. We introduce the "neural state vector" for N neurons in a network as
I~(t) >=
L p[n, t]l n >,
{fl}
(5)
273
Stochastic Dynamics of Three-State Neural Networks
where the sum is taken over all possible network states.
With these definitions, we can write the master equation for a network with the
transition rates shown in Figure 1, using the outer-product representations of operators (Sakurai, 1985). For example:
lai >< q;j =
( 0o 0 1)
0
0
0
0
0
(6)
.
?
The master equation then takes the form of an evolution equation:
,
8
- 8t 1<I>(t) >= LI<I>(t) >
with the network "Liouvillian" L given by:
(7)
N
N
1 N
L = a I)lai}(ad -Iri}(a;j) + L)h}(1'il-l a i}(,'d)e2(=
wijlaj}(ajl)
.
.
n .
1=1
.=1
)=1
L
+,6L:~1 (Iri}(ril-Iqi}(r;!) + L:~l (lqi}(qil-lai}(qiI)OlUi L:7=1 wijlaj}(ajl).(8)
where n is an average number of connections to each neuron, and Wij is the "weight"
from the jth to the ith neuron. Thus the weights are normalized with respect to
the average number n of connections per neuron.
The master equation given here is the same as the one introduced by Cowan using
Gell-Mann matrices (Cowan, 1991). However, we note that with the outer-product
representation, we can extend the description from two to three-state neurons simply by including one more basis vector.
In analogy with the analysis of two-state neurons, we introduce the state vector:
N
(a r cJI =
IT (qi(q;! + ri(rd + ai(ail).
(9)
i=1
where the product is taken as a direct product, and ai, ri, and qi are parameters.
We also introduce the point moments ?ai(t)>>, ?qi(t)>>, and ?ri(t)>> as the
probability that the ith neuron is active, quiescent, and refract.ory respectively, at
time t. Similarly, we can define the multiple moment, for example, ?aiqp'k .. , (t)>>
as the probability that the ith neuron is active, the jth neuron is quiescent, the kth
neuron is refractory and so on at time t. Then, it can be shown that they are given
by:
= (a = r = q= 1lsi}(sd @ ISj)(Sjl @ ISk)(Skl ... 1<I>(t)},
?SiSjSk ... (t)>>
S
(10)
= a, " ,q
For example,
?riqjak(t)>>
= (a = r = q = 1I"i}(ril @ Iqj}(qjl @ lak)(akl<I>(t)}
(11)
We note the following relations,
?ai(t)>> + ?qi(t)>> + ?ri(t)>>
=1
(12)
and
?aHt)>>
= ?ai(t)>>,
?r;(t)>>
= ?"i(t)>>,
?q;(t)>>
= ?qi(t)>>,
(13)
Toru Ohira, Jack D. Cowan
274
3
THE HIERARCHY OF MOMENT EQUATIONS
We can now obtain an equation of motion for the moments. As is typical in the
case of many-body problems, we obtain an analogue of the BBGKY hierarchy of
equations (Doi, 1976b). This can be done by using the definition of moments, the
master equation, and the a-r-q state vector. We show the hierarchy up to the second
order:
a
--at ?airj? =
-a( ?aiaj? - ?airj?) + (3?ai 1'j?
1
N
+ ?air j(}2(=n~
'"' Wikak)>>
k=l
-?rirj(}2(~'E~=1 Wikak)>> -
?qjrj(}lUf'E:=l Wikak)>> (19)
We note that since
(20)
?ai? + ?ri? + ?qi? = 1,
one of the parameters can be eliminated. We also note that the equations are
coupled into higher orders in this hierarchy. This leads to a need for approximation
schemes which can terminate the hierarchy at an appropriate order.
In the following, we introduce first and the second moment level approximation
schemes. For simplicity, we consider the special case in which (}l and (}2 are linear
and equal.
With the above simplication the first moment (mean field) approximation leads to:
Stochastic Dynamics of Three-State Neural Networks
-:t
?ai?
= a?ai? -
o
275
Wi( ?ri? + ~qi?)
(22)
j3_
- Ot ?ri? = -a?ai? + ?ri? + Wi?ri?,
o
- Ot ?qi?
_
= - j3?ri? + Wi?qi?,
(23)
(24)
where
1 N
WI
= iiL wlk?ak?,
(25)
k=l
We also obtain the second moment approximation as:
o
- Ot ?ai?
= a?ai? -
1 N
ff~ Wij( ?qiaj? + ?riaj?),
(26)
)=1
o
1 N
- Ot ?ri? = -a?ai? + j3?ri? + ff~ Wij?riaj?,
(27)
)=1
o
- Ot ?qi? = -j3?ri?
1 N
+ ff~ Wij?qiaj?,
(28)
)=1
-:t
o
+ ?qiaj?)
-Wji( ?airj? + ?aiqj?),
?aiaj? = 2a?aiaj? - Wij( ?riaj?
- Ot ?airj?
= -a( ?aiaj? -
(29)
_
?airj?) + j3?air j? + Wji?airj?
-Wij(<<7'i7'j?
+ ?qi7'j?),
(31)
where
(32)
We note that the first moment dynamics obtained via the first approximation differs
from that obtained from the second moment approximation, In the next section,
we briefly examine this difference by comparing these approximations with Monte
Carlo simulations.
276
4
Toru Ohira, Jack D. Cowan
COMPARISON WITH SIMULATIONS
In this section, we compare first and second moment approximations with Monte
Carlo simulation of a one dimensional ring of three-state neurons. This was studied
in a previous publication (Ohira and Cowan, 1993) for two-state neurons. As shown
there, each three-state neuron in the ring interacts with its two neighbors.
More precisely, the Liouville operator is
N
N
L = a 2:)lai}(ail-/ri}(a;!) + f3 L(/ri}{ril-/qi}(ril)
i=1
i=1
1
N
+2W2 L(/ri}(ri/-l a i}(r;I)(/ai+l}(ai+d + /ai-l}{ai-d)
i=1
1
N
+2W1 L(/qi}{q;/ - /ai}(qil)(/ai+l}{ai+d + /ai-J}(ai-d)
i=1
(33)
We now define the dynamical variables of interest as follows:
Xa
1
1 1 1
=N
L
i=1
?aj?,
Xr
=N
L ?ri?, Xq
=N
;=1
N
1]aa = N L ?ai a i+l?, 1]rr
;=1
1
L ?qi?,
(34)
;=1
N
= N L ?ri r i+l?, 1]ar
i=1
1
N
=NL
?ai r i+l?.
(35)
i=1
Then, for this network, the first moment approximation is given by
8
- 8tXa
8
- at Xr
Xq
-
ax - W2XaXr - WIXqXa,
=
-ax - f3Xr
=
1 - Xa - Xr.
+ W2XqXa,
(36)
The second moment approximation is given by
8
- at Xa
=
8
- 8t Xr
=
{)
+ 1) -
- 8t1]aa
-
2a1]aa - W21]ar(Xa
8
- 8t1]ar
=
-a('1aa -1]ar) - f31]ar
Wl(Xa
+ 1)(Xa -1]ar -1]aa),
1
+ 2 W2 '1ar(Xa + 1),
1
+21]rrXa + WIXa(Xr -1]rr -1]ar)'
(37)
277
Stochastic Dynamics of Three-State Neural Networks
Monte Carlo simulations of a ring of 10000 neurons were performed and compared
with the first and second moment approximation predictions. We fixed the following
parameters:
(38)
a = 1.0, f3 = 0.2, Wl = 0.01? WO, W2 = 0.6? Wo
x",. ?.... - .................. __ .............. - ..... ..
x. I .'
"
--------_ ....... -- .. _--x, ?.?
x" .?
X.'"
(A)
- -- ..... -- .... -_ .. ------_.
., ... ..
.. ........... _----_ ........ _----
(8)
Figure 2: Comparison of Monte Carlo simulations (dots) with the first moment
(dashed line) and the second moment (solid line) approximations for the three state
case with the fraction of total active and refractory state variables Xa (A) and Xr
(B). Each graph is labeled by the values of wo/a.
We varied Wo and sampled the numerical dynamics of these parameters. Some
comparisons are shown in Figure 2 for the time dependence of the total number
of active and refractory state variables. We dearly see the improvement of the
second over the first moment level approximation. More simulations with different
parameter ranges remain to be explored.
5
CONCLUSION
We have introduced here a neural network master equation using the outer-product
representation. In this representation, the extension from two to three-state neurons is transparent. We have taken advantage of this natural extension to analyse
three-state networks. Even though the calculations involved are more intricate, we
278
Torn Ohira, Jack D. Cowan
have obtained results indicating that the second moment level approximation is significantly more accurate than the first moment level approximation. We also note
that as in the two-state case, the first moment level approximation produces more
activation than the simulation. FUrther analytical and theoretical investigations
are needed to fully uncover the dynamics of three-state networks described by the
master equation introduced above.
Acknowledgements
This work was supported in part by the Robert R. McCOlmick fellowship at the
University of Chicago, and in part by grant No. N0014-89-J-I099 from the US
Department of the Navy, Office of Naval Research.
References
Cowan JD (1991) Stochastic neurodynamics in Advances in Neural Information
Processing Systems (D. S. Touretzky, R. P. Lippman, J. E. Moody, ed.), vol. 3,
Morgan Kaufmann Publishers, San Mateo
Doi M (1976a) Second quantization representation for classical many-particle system. J. Phys. A: Math Gen. 9:1465-1477.
Doi M (1976b) Stochastic theory of diffusion-controlled reactions. J. Phys. A: Math.
Gen. 9:1479.
Grassberger P, Scheunert M (1980) Fock-space methods for identical classical objects. Fortschritte der Physik 28:547
Meunier C, Hansel D, Verga A (1989) Information processing in three-state neural
networks. J. Stat. Phys. 55:859
Ohira T, Cowan JD (1993) Master-equation approach to stochastic neurodynamics.
Phys. Rev. E 48:2259
Ohira T, Cowan JD (1994) Feynman Diagrams for Stochastic Neurodynamics. In
Proceedings of Fifth Australian Conference of Neural Networks, pp218-221
Sakurai JJ (1985) Modern Quantum Mechanics. Benjamin/Cummings, Menlo Park
| 991 |@word normalized:1 briefly:1 classical:3 australian:1 evolution:1 liouville:1 ril:4 physik:1 tokyo:1 laboratory:1 simulation:9 stochastic:11 dependence:1 interacts:1 kth:1 mann:1 solid:1 moment:21 outer:5 configuration:1 transparent:1 investigation:1 past:1 reaction:1 extension:3 current:3 comparing:1 motion:1 liouvillian:1 reformulate:1 activation:1 iil:1 jack:5 grassberger:2 numerical:1 chicago:3 robert:1 refractory:4 fh:1 jp:1 extend:2 item:1 hansel:2 neuron:26 ai:24 ith:4 wl:2 rd:1 finite:1 mathematics:1 similarly:1 refract:1 particle:1 math:2 dot:1 varied:1 rather:1 direct:2 publication:1 office:1 gotanda:1 introduced:4 derived:1 ax:2 naval:1 improvement:1 connection:2 modelling:1 ajl:2 introduce:5 scheunert:2 dqi:1 intricate:1 behavior:1 examine:1 mechanic:1 der:1 wji:2 morgan:1 dynamical:1 employed:1 csl:1 relation:1 increasing:1 wij:6 dashed:1 including:1 notation:2 multiple:1 adai:1 analogue:1 denoted:1 natural:1 ail:2 akl:1 special:1 calculation:1 equal:1 finding:1 field:1 f3:2 lai:5 eliminated:1 verga:2 identical:1 qi:14 park:1 j3:4 basic:1 a1:1 prediction:1 controlled:1 coupled:1 represent:1 xq:2 grant:1 modern:1 acknowledgement:1 composed:1 t1:2 iqi:2 fellowship:1 fully:1 sd:1 diagram:1 macroscopic:1 publisher:1 ak:1 ot:6 w2:3 analogy:2 interest:1 cowan:15 studied:1 mateo:1 co:1 qii:1 nl:1 limited:1 ohira:10 activated:1 range:1 equa:1 accurate:1 supported:1 asynchronous:1 jth:2 differs:1 inner:1 uchicago:1 xr:6 lippman:1 neighbor:1 continuing:1 i7:1 fifth:1 f31:1 significantly:1 theoretical:1 cji:1 transition:7 formalism:2 wo:4 sjl:1 markovian:1 quantum:1 ar:8 san:1 sakurai:2 operator:3 jj:1 influence:1 ory:1 iqj:1 iri:3 active:4 assumed:1 simplicity:1 quiescent:3 torn:1 qjl:1 lsi:1 wlk:1 neurology:1 scheme:5 per:1 neurodynamics:4 gell:1 aiaj:4 terminate:1 write:1 lak:1 menlo:1 vol:1 moody:1 hierarchy:6 w1:1 four:1 us:1 diffusion:1 noise:1 graph:1 li:1 japan:1 qil:3 labeled:1 fraction:1 sum:1 body:2 skl:1 higashi:1 master:13 site:2 ff:3 cycle:1 ad:2 performed:2 meunier:2 benjamin:1 fl:1 dynamic:7 depend:1 activity:1 air:2 il:2 spin:1 precisely:1 kaufmann:1 explored:1 ri:20 correspond:1 basis:2 easily:1 aht:1 essential:1 quantization:2 represented:1 produced:1 carlo:6 w21:1 department:1 monte:6 doi:4 depts:1 fortschritte:1 phys:4 smoothly:1 touretzky:1 airj:6 ed:1 navy:1 definition:2 remain:1 simply:1 isj:1 wi:4 say:1 rev:1 involved:1 e2:1 toru:4 analyse:1 sampled:1 aa:5 taken:3 advantage:1 rr:2 equation:21 analytical:1 dearly:1 needed:1 product:9 uncover:1 sony:2 labelled:1 feynman:1 higher:1 gen:2 cummings:1 typical:1 lqi:1 apply:1 synapse:1 done:1 description:2 though:1 appropriate:1 dirac:1 xa:8 total:2 la:1 indicating:1 jd:3 produce:1 ring:3 object:1 tions:1 iq:1 stat:1 aj:2 isk:1 |
7,234 | 992 | Grouping Components of
?
Three-Dimensional Moving Objects In
Area MST of Visual Cortex
Richard S. Zemel
Carnegie Mellon University
Department of Psychology
Pittsburgh, PA 15213
zemel?lcmu. edu
Terrence J. Sejnowski
CNL, The Salk Institute
P.O. Box 85800
San Diego, CA 92186-5800
terry?lsalk.edu
Abstract
Many cells in the dorsal part of the medial superior temporal
(MST) area of visual cortex respond selectively to spiral flow
patterns-specific combinations of expansion/ contraction and rotation motions. Previous investigators have suggested that these
cells may represent self-motion. Spiral patterns can also be generated by the relative motion of the observer and a particular object.
An MST cell may then account for some portion of the complex
flow field, and the set of active cells could encode the entire flow;
in this manner, MST effectively segments moving objects. Such
a grouping operation is essential in interpreting scenes containing
several independent moving objects and observer motion. We describe a model based on the hypothesis that the selective tuning
of MST cells reflects the grouping of object components undergoing coherent motion. Inputs to the model were generated from
sequences of ray-traced images that simulated realistic motion situations, combining observer motion, eye movements, and independent object motion. The input representation was modeled after
response properties of neurons in area MT, which provides the primary input to area MST. After applying an unsupervised learning
algorithm, the units became tuned to patterns signaling coherent
motion. The results match many of the known properties of MST
cells and are consistent with recent studies indicating that these
cells process 3-D object motion information.
166
1
Richard S. Zemel, Terrence J. Sejnowski
INTRODUCTION
A number of studies have described neurons in the dorsal part of the medial superior
temporal (MSTd) monkey cortex that respond best to large expanding/contracting,
rotating, or shifting patterns (Tanaka et al., 1986; Duffy & Wurtz, 1991a). Recently
Graziano et al. (1994) found that MSTd cell responses correspond to a point in a
multidimensional space of spiral motions, where the dimensions are these motion
types.
Combinations of these motions are generated as an animal moves through its environment, which suggests that area MSTd could playa role in optical flow analysis.
When an observer moves through a static environment, a singularity in the flow
field known as the focus of expansion may be used to determine the direction of
heading (Gibson, 1950; Warren & Hannon, 1988). Previous computational models
of MSTd (Lappe & Rauschecker, 1993; Perrone & Stone, 1994) have shown how
navigational information related to heading may be encoded by these cells. These
investigators propose that each MSTd cell represents a potential heading direction
and responds to the aspects of the flow that are consistent with that direction.
In natural environments, however, MSTd cells are often faced with complex flow
patterns produced by the combination of observer motion with other independentlymoving objects. These complex flow fields are not a single spiral pattern, but instead
are composed of multiple spiral patterns. This observation that spiral flows are local
subpatterns in flow fields suggests that an MSTd cell represents a particular regular
subpattern which corresponds to the aspects of the flow field arising from a single
cause-the relative motion of the observer and some object or surface in the scene.
Adopting this view implies a new goal for MST: the set of MST responses accounts
for the flow field based on the ensemble of motion causes.
An MST cell that responds to a local subpattern accounts for a portion of the flow
field, specifically the portion that arises from a single motion cause. In this manner,
MST can be considered to be segmenting motion signals. As in earlier models, the
MSTd cell responds to the aspects of flow consistent with its motion hypothesis,
but here a cell's motion hypothesis is not a heading direction but instead represents
the 3-D motion of a scene element relative to the observer. This encoding may be
useful not only in robustly estimating heading detection, but may also facilitate
several other tasks thought to occur further down the motion processing stream,
such as localizing objects and parsing scenes containing multiple moving objects.
In this paper we describe a computational model based on the hypothesis that an
MST cell signals those aspects of the flow that arise from a common underlying
cause. We demonstrate how such a model can develop response properties from the
statistics of natural flow images, such as could be extracted from MT signals, and
show that this model is able to capture several known properties of information
processing in MST.
2
THE MODEL
The input to the system is a movie containing some combination of observer motion,
eye movements, and a few objects undergoing independent 3-D motion. An optical
Grouping Components of 3-D Moving Objects in Area MST of Visual Cortex
167
flow algorithm is then applied to yield local motion estimates; this flow field is the
input to the network, which consists of three layers. The first layer is designed after
monkey area MT. The connectivity between this layer and the second layer is based
on MST receptive field properties, and the second layer has the same connectivity
pattern to the output layer. The weights on all connections are determined by a
training algorithm which attempts to force the network to recreate the input pattern
on the output units. We discuss the inputs, the network, and the training algorithm
in more detail below.
2.1
STIMULI
The flow field input to the network is produced from a movie. The various movies
are intended to simulate natural motion situations. Sample situations include one
where all motion is due to the observer's movement, and the gaze is in the motion
direction. Another situation that produces a qualitatively different flow field is when
the gaze is not in the motion direction. A third situation includes independent
motion of some of the objects in the environment. Each movie is a sequence of
images that simulates one of these situations.
The images are created using a flexible ray-tracing program, which allows the simulation of many different objects, backgrounds, observer/camera motions, and lighting effects. We currently employ a database of 6 objects (a block of swiss cheese,
a snail, a panther, a fish, a ball, and a teapot) and three different backgrounds. A
movie is generated by randomly selecting one to four of the objects, and a background. To simulate one of the motion situations, a random selection of motion
parameters follows: a). The observer's motion along (x, z) describes her walking;
b). The eyes can rotate in (x, y), simulating the tracking of an object during motion;
c). Each object can undergo independent 3-D motion. A sequence of 15 images is
produced by randomly selecting 3-D initial positions and then updating the pose
of the camera and each object in the image based on these motion parameters.
Figure 1 shows 3 images selected from a movie generated in this manner.
We apply a standard optical flow technique to extract a single flow field from each
synthetic image sequence. Nagel's (1987) flow algorithm is a gradient-based scheme
which performs spatiotemporal smoothing on the set of images and then uses a
multi-resolution second-order derivative technique, in combination with an oriented
smoothness relaxation scheme, to produce the flow field.
2.2
MODEL INPUT AND ARCHITECTURE
The network input is a population encoding of these optical flow vectors at each
location in a 21x31 array by small sets of neurons that share the same receptive
field position but are tuned to different directions of motion. The values for each
input unit is computed by projecting the local flow vector onto the cell's preferred
direction. We are currently using 4 inputs per location, with evenly spaced preferred
directions and a tuning half-width of 45?.
This popUlation encoding in the input layer is intended to model the response of
cells in area MT to a motion sequence. The receptive field (RF) of each model MT
unit is determined by the degree of spatial smoothing and subsampling in the flow
| 992 |@word effect:1 implies:1 subpattern:2 direction:9 move:2 receptive:3 simulation:1 primary:1 contraction:1 responds:3 during:1 self:1 width:1 gradient:1 simulated:1 initial:1 evenly:1 stone:1 selecting:2 tuned:2 demonstrate:1 singularity:1 performs:1 motion:39 interpreting:1 modeled:1 considered:1 image:9 recently:1 parsing:1 superior:2 rotation:1 mst:15 realistic:1 common:1 mt:5 designed:1 medial:2 half:1 selected:1 currently:2 mellon:1 neuron:3 observation:1 smoothness:1 tuning:2 reflects:1 situation:7 provides:1 location:2 moving:5 cortex:4 surface:1 along:1 playa:1 encode:1 consists:1 focus:1 recent:1 connection:1 ray:2 rauschecker:1 manner:3 coherent:2 tanaka:1 multi:1 able:1 suggested:1 below:1 pattern:9 entire:1 her:1 determine:1 selective:1 program:1 signal:3 estimating:1 underlying:1 multiple:2 navigational:1 rf:1 flexible:1 terry:1 shifting:1 natural:3 force:1 match:1 animal:1 monkey:2 smoothing:2 spatial:1 field:15 scheme:2 movie:6 eye:3 temporal:2 teapot:1 represents:3 multidimensional:1 unsupervised:1 created:1 wurtz:1 extract:1 stimulus:1 represent:1 richard:2 unit:4 employ:1 few:1 randomly:2 oriented:1 composed:1 segmenting:1 adopting:1 cell:18 background:3 local:4 relative:3 contracting:1 intended:2 encoding:3 attempt:1 detection:1 nagel:1 subpatterns:1 undergo:1 simulates:1 degree:1 flow:27 consistent:3 suggests:2 snail:1 share:1 spiral:6 camera:2 psychology:1 architecture:1 heading:5 block:1 warren:1 swiss:1 lappe:1 signaling:1 institute:1 area:8 rotating:1 gibson:1 recreate:1 tracing:1 thought:1 dimension:1 regular:1 earlier:1 qualitatively:1 san:1 onto:1 selection:1 localizing:1 cause:4 applying:1 useful:1 preferred:2 cheese:1 active:1 resolution:1 pittsburgh:1 spatiotemporal:1 synthetic:1 fish:1 array:1 arising:1 per:1 population:2 terrence:2 carnegie:1 ca:1 graziano:1 gaze:2 expanding:1 diego:1 connectivity:2 four:1 expansion:2 complex:3 traced:1 us:1 containing:3 hypothesis:4 pa:1 element:1 walking:1 updating:1 derivative:1 arise:1 relaxation:1 database:1 account:3 potential:1 role:1 respond:2 capture:1 includes:1 salk:1 stream:1 movement:3 position:2 view:1 observer:11 x31:1 environment:4 portion:3 layer:7 third:1 down:1 specific:1 segment:1 occur:1 became:1 undergoing:2 ensemble:1 correspond:1 yield:1 spaced:1 mstd:8 scene:4 grouping:4 essential:1 aspect:4 simulate:2 panther:1 various:1 effectively:1 produced:3 optical:4 lighting:1 faced:1 describe:2 department:1 sejnowski:2 duffy:1 zemel:3 combination:5 perrone:1 ball:1 describes:1 encoded:1 cnl:1 visual:3 tracking:1 statistic:1 projecting:1 static:1 corresponds:1 extracted:1 sequence:5 discus:1 goal:1 propose:1 combining:1 operation:1 specifically:1 determined:2 response:5 apply:1 box:1 simulating:1 robustly:1 indicating:1 selectively:1 produce:2 include:1 hannon:1 object:20 subsampling:1 arises:1 develop:1 dorsal:2 pose:1 investigator:2 rotate:1 facilitate:1 |
7,235 | 993 | Visual Speech Recognition with
Stochastic Networks
Javier R. Movellan
Department of Cognitive Science
University of California San Diego
La Jolla, Ca 92093-0515
Abstract
This paper presents ongoing work on a speaker independent visual
speech recognition system. The work presented here builds on previous
research efforts in this area and explores the potential use of simple
hidden Markov models for limited vocabulary, speaker independent
visual speech recognition. The task at hand is recognition of the first
four English digits, a task with possible applications in car-phone
dialing. The images were modeled as mixtures of independent
Gaussian distributions, and the temporal dependencies were captured
with standard left-to-right hidden Markov models. The results indicate
that simple hidden Markov models may be used to successfully
recognize relatively unprocessed image sequences. The system achieved
performance levels equivalent to untrained humans when asked to
recognize the fIrst four English digits.
1 INTRODUCTION
Visual articulation is an important source of information in face to face speech perception.
Laboratory studies have shown that visual information allows subjects to tolerate an extra
4-dB of noise in the acoustic signal. This is particularly important considering that each
decibel of signal to noise ratio translates into a 10-15% error reduction in the
intelligibility of entire sentences (McCleod and SummerfIeld, 1990). Lip reading alone
provides a basis for understanding for a large majority of the hearing impaired and when
supplemented by acoustic or electrical signals it allows fluent understanding of speech in
highly trained subjects. However visual information plays more than a simple
compensatory role in speech perception. From early on humans are predisposed to
integrate acoustic and visual information. Sensitivity to correspondences in auditory and
visual information for speech events has been shown in 4 month old infants (Spelke,
1976; Kuhl & Meltzoff, 1982). By 6 years of age, humans consistently use audio visual
contingencies to understand speech (Massaro, 1987). By adulthood, visual articulation
automatically modulates perception of the acoustic signal. Under laboratory conditions it
is possible to create powerful illusions in which subjects mistakenly hear sounds which
are biased by visual articulations. Subjects in these experiments are typically unaware cf
852
Javier Movellan
the discrepancy between the visual and auditory tracks and their experience is that of a
unified auditory percept (McGurk & McDonnald, 1976).
Recent years have seen a revival of interest in audiovisual speech perception both in
psychology and in the pattern recognition literature. There have been isolated efforts to
build synthetic models of visual and audio-visual speech recognition (Petahan, 1985;
Nishida, 1986; Yuhas, Goldstein, Sejnowski & Jenkins, 1988; Bregler, Manke, Hild &
Waibel, 1993; Wolff, Prassad, Stork, & Hennecke, 1994). The main goal of these efforts
has been to explore different architectures and visual processing techniques and to
illustrate the potential use of visual information to improve the robustness of current
speech recognition systems. Cognitive psychologists have also developed high level
models of audio-visual speech perception that describe regularities in the way humans
integrate visual and acoustic information (Massaro, 1987). In general these studies
support the idea that human responses to visual and acoustic stimuli are conditional
independent. This regularity has been used in some synthetic systems to simplify the
task of integrating visual and acoustic signals (Wolff, Prassad, Stork, & Hennecke,
1994). Overall, multimodal speech perception is still an emerging field in which a lot cf
exploration needs to be done. The work presented here builds on the previous research
efforts in this area and explores the potential use of simple hidden Markov models 1ir
limited vocabulary, speaker independent visual speech recognition. The task at hand is
recognition of the first four English digits, a task with possible applications in car-phone
dialing.
2 TRAINING SAMPLE
The training sample consisted of 96 digitized movies of 12 undergraduate students (9
males,3 females) from the Cognitive Science Department at UCSD. Video capturing was
performed in a windowless room at the Center for Research in Language at UCSD.
Subjects were asked to talk into a video camera and to say the first four digits in English
twice. Subjects could monitor the digitized images in a small display conveniently
located in front of them. They were asked to position themselves so that that their lips be
roughly centered in the feed-back display. Gray scale video images were digitized at 30
fps, 100x75 pixels, 8 bits per pixel. The video tracks were hand segmented by selecting a
few relevant frames before and after the beginning and end of activity in the acoustic track.
Statistics of the entire training sample are shown in table 1.
Table 1: Frame number statistics.
Digit
"One"
"Two"
"Three"
"Four"
Average
8.9
9.6
9.7
10.6
S.D.
2.1
2.1
2.3
2.2
3 IMAGE PREPROCESSING
There are two different approaches to visual preprocessing in the visual speech recognition
literature (Bregler, Manke, Hild & Waibel, 1993). The first approach, represented by the
work of Wolff and colleagues (Wolff, Prassad, Stork, & Hennecke, 1994) favors
sophisticated image preprocessing techniques to extract a limited set of hand-crafted
features (e.g., height and width of the lips). The advantage of this approach is that it
Visual Speech Recognition with Stochastic Networks
853
drastically reduces the number of input dimensions. This translates into lower variability
of the signal, potentially improved generalization, and large savings in computing time.
The disadvantage is that vital information may be lost when compressing the image into
a limited set of hand-crafted features. Variability is reduced at the possible expense cf
bias. Moreover, tests have shown that subtle holistic features such as the wrinkling and
protrusion of the lips may play an important role in human lip-reading (Montgomery &
Jackson, 1983). The second approach to visual preprocessing emphasizes preserving the
original images as much as possible and letting the recognition engine discover the
relevant features in the images. In most cases, images are low-pass filtered and
dimension-reduced by using principal component analysis. The results in this papers
indicate that good results can be obtained even without the use of principal components.
In this investigation image preprocessing consisted of the following phases:
1. Symmetry enforcement: At each time frame the raw images were symmetrized by
averaging pixel by pixel the left and right side of each image, using the vertical midline
as the axis of symmetry. For convenience from now on we will refer to the raw images as
"rho-images" and the symmetrized images as "sigma-images." The potential benefits cf
sigma-images are robustness, and compression, since the number of relevant pixels is
reduced by half.
2. Temporal differentiaion: At each time frame we calculated the pixel by pixel differences
between present sigma-images and immediately past sigma-images. For convenience we
refer to the resulting images as "delta-images." One of the potential advantages of deltaimages in the visual domain is their robustness to changes in illumination and the fuct
that they emphasize the dynamic aspects of the visual track.
3. Low pass filtering and subsampling: The sigma and delta images were compressed and
subsampled using 20x 15 equidistant Gaussian filters. Different values of the standard
deviation of the Gaussian filters were tested.
4. Logistic thresholding and scaling: The sigma and delta images were independently
thresholded by feeding the output of the Gaussian filters through a according to the
following equation
Y = 256 j(K
7r
r;;
-v3a
(x -
J.L))
where f is the logistic function, and J.L, a, are respectively the average and standard
deviation of the gray level distribution of entire image sequences. The constant K
controls the sharpness of the logistic function. Assuming an approximately Gaussian
distribution of gray levels when K=1 the thresholding function approximates histogram
equalization, a standard technique in visual processing. Three different K values were
tried: 0.3, 0.6 and 1.2.
5. Composites of the relevant portions of the blurred sigma and delta images were fed to
the recognition network. The number of pixels of each processed image was 300 (150
from the blurred sigma images and 150 from the blurred delta images). Figure 1 shows
the effect of the different preprocessing stages.
854
Javier Movellan
Figure 1: Image Preprocessing. 1) Rho-Image. 2) Sigma-Image.
3) Delta-Image. 4) Filtered and Sharpened Composite.
4 RECOGNITION NETWORK
We used the standard approach in limited vocabulary systems: a bank of hidden Markov
models, one per word category, independently trained on the corresponding word
categories. The images were modeled as mixtures of continuous probability distributions
in pixel space. We tried mixtures of Gaussians and mixtures of Cauchy distributions.
The mixtures of Cauchy distributions were very stable numerically but they did perform
very poorly when compared to the Gaussian mixtures. We believe the reason for their
poor performance is the tendency of Cauchy-based maximum-likelihood estimates to
focus on individual exemplars. Gaussian-based estimates are much more prone to blend
exemplars that belong to the same cluster. The initial state probabilities, transition
probabilities, mixture coefficients, mixture centroids and variance parameters were trained
using the E-M algorithm.
We initially encountered severe numerical underflow problems when using the E-M
algorithm with Gaussian mixtures. These instabilities were due to the fact that the
probability densities of images rapidly went to zero due to the large dimensionality of the
images. Trimming the outputs of the Gaussian and using very small Gaussian gains did
not work well. We solved the numerical problems in the following way: 1) Constraining
all the variance parameters for all the states and mixtures to be equal. This allowed
pulling out a constant in the likelihood-function of the mixtures, avoiding most
numerical problems. 2) Initializing the mixture centroids using linear segmentation
followed by the K-means clustering algorithm. For example, ifthere were 4 visual frames
and 2 states, the first 2 frames were assigned to state 1 and the last 2 frames to state 2. Kmeans was then used independently on each of the states and their assigned frames. This
is a standard initialization method in the acoustic domain (Rabiner & Bing-Hwang,
1993). Since K-means can be trapped in local minima, the algorithm was repeated 20
times with different starting point and the best solution was fed as the starting point for
the E-M algorithm.
5 RESULTS
The main purpose of this study was to fmd simple image preprocessing techniques that
would work well with hidden Markov models. We tested a wide variety of architectures
and preprocessing parameters. In all cases the results were evaluated in terms cf
generalization to new speakers. Since the training sample is small, generalization
performance was estimated using the jackknife procedure. Models were trained with 11
Visual Speech Recognition with Stochastic Networks
855
subjects, leaving one subject out for generalization testing. The entire procedure was
repeated 12 times, each time leaving a different subject out for testing. Results are thus
based on 96 generalization trials (4 digits x 12 subjects x 2 observations per subject). In
all cases we tested several preprocessing techniques using 20 different architectures with
different number of states (1,3,5,7,9) and mixtures per state (1,3,5,7). To compare the
effect of each processing technique we used the average generalization performance of the
best 4 architectures out of the 20 architectures tested.
85
65
45
25 +---L._ _
Rho Sigma Delta+Sigma
Figure 2: Average performance with the rho, sigma, and delta images.
Figure 2 shows the effects of symmetry enforcement and temporal differentiation.
Symmetry enforcement had the benefit of reducing the input dimensionality by half and,
as the figure show it did not hinder recognition performance. Using delta images had a
very positive effect on recognition performance, as the figure shows. Figure 3 shows the
effect of varying the thresholding constant and the standard deviation of the Gaussian
filters. Best performance was obtained with blurring windows about 4 pixel wide and
with thresholding just about histogram equalization.
84
83
82
81
---'-K=i.2
---K=O.6
K=O.3
80
3 4 5
Stamdard Deviation of
Gaussian Filter.
Figure 3: Effect of blurring and sharpening.
856
Javier Movellan
Table 2 shows the effects of variations in the number of states (S) and Gaussian mixtures
(G) per state. The number within each cell is the percentage of simulations for which a
particular combination of states and mixtures performed best out of the 20 architectures
tested.
Table 2: Effect of varying the the number of states (S) and Gaussian mixtures (G).
SI
S3
S5
S7
S9
Gl
0.00%
0.00%
3.12%
6.25%
6.25%
G3
0.00%
21.87%
9.37%
12.5%
0.00%
G5
0.00%
12.5%
15.62%
0.00%
3.12%
G7
0.00%
6.25%
0.00%
3.12%
0.00%
Best overall performance was obtained with about 3 states and 3 mixtures per state.
Peak performance was also obtained with a 3-state, 3-mixture per state network, with a
generalization rate of 89.58% correct.
To compare these results with human performance, 9 subjects were tested on the same
sample. Six subjects were normal hearing adults who were not trained in lip-reading.
Three were hearing impaired with profound hearing loss and had received training in lip
reading at 2 to 8 years of age. The mean correct response for normal subjects was 89.93
% correct, just about the same rate as the best artificial network. The hearing impaired
had an average performance of 95.49% correct, significantly better than our network.
Table 3: Confusion matrix of the best artificial system.
"One"
"Two"
"Three"
"Four"
1
100.00%
4.17%
12.5%
8.33%
2
0.00%
87.50%
0.00%
4.17%
3
0.00%
4.17%
83.33%
0.00%
4
0.00%
4.17%
4.17%
87.50%
Table 4: Average human confusion matrix.
"One"
"Two"
"Three"
"Four"
1
89.36%
1.39%
9.25%
4.17%
2
0.46%
98.61%
3.24%
0.46%
3
8.33%
0.00%
85.64%
1.85%
4
1.85%
0.00%
1.87%
93.52%
Tables 3 and 4 show the confusion matrices for the best network and the average
confusion matrix with all 9 subjects combined. The correlation between these two
matrices was 0.99. This means that 98% of the variance in human confusions can be
accounted for by the artificial model. This suggests that the representational space
learned by the artificial system may be a reasonable model of the representational space
used by humans. Figure 5 shows the representations learned by a network with 6 states
and 1 mixture per state. Each column is a different digit, starting with "one." Each row
Visual Speech Recognition with Stochastic Networks
857
is a different temporal state. The two pictures within each cell are sigma and delta image
centroids. As the figure shows, the identity of individual exemplars is lost but the
underlying dynamics of the digits are preserved. The digits can be easily recognized
when played as a movie.
Figure 4: Dynamic representations learned by a simple network.
6
CONCLUSIONS
This paper shows that simple stochastic networks, like hidden Markov models, can be
successfully applied for visual speech recognition using relatively unprocessed images.
The performance level obtained with these networks roughly matches untrained human
performance. Moreover, the representational space learned by these networks may be a
reasonable model of the representations used by humans. More research should be done to
better understand how humans integrate visual and acoustic information in speech
perception and to develop practical models for robust audio-visual speech recognition.
References
Bregler c., Manke S., Hild H. & Waibel A. (1993) Bimodal Sensor Integration on the
Example of "Speech-Reading". Proc ICNN-93, 11,667-677.
Kuhl P. & Meltzoff A (1982) The Bimodal Development of Speech in Infancy. Science,
218, 1138-1141.
MacLeod A. & Summerfield Q. (1990) A Procedure for Measuring Auditory and Audiovisual Speech-Reception Measuring Thresholds for Sentences in Noise: Rationale,
Evaluation and Recommendations for Use. British Journal of Audiology. 24,29-43.
Massaro D. (1987) Speech Perception by Ear and Eye. In Dodd B. & Campbell R. (ed.)
Hearing by Eye: The Psychology of Lip-Reading. London, LEA, 53-83.
Massaro D., Cohen M & Getsi (1993) Long-Term Training, Transfer and Retention in
Learning to Lip-read. Perception and Psychophysics, 53,549-562.
858
Javier Movellan
McGurk H. & MacDonald J. (1976) Hearing Lips and Seeing Voices. Nature, 264, 126130.
Montgomery A. & Jackson P. (1983) Physical Characteristics of the Lips Underlying
Vowel Lipreading Performance. Journal of the Acoustical Society of America, 73,21342144.
Nishida S. (1986) Speech Recognition Enhancement by Lip Information Proceedings of
ACMlCHI86, 198-204.
Petajan E. (1985) Automatic Lip Reading to Enhance Speech Recognition. IEEE CVPR
85, 40-47.
Rabiner L., Bing-Hwang J. (1993) Fundamentals of Speech Recognition. New Jersey,
Prentice Hall.
Spelke E. (1976)
533-560.
Infant's Intermodal Perception of Events. Cognitive Psychology, 8,
Yuhas B., Goldstein T., Sejnowski T., Jenkins R. (1988) Neural Network Models cf
Sensory Integration for Improved Vowel Recognition. Proceedings IEEE 78, 16551668.
Wolff G., Prassad L., Stork D., Hennecke M. (1994) Lipreading by Neural Networks:
Visual Preprocessing, Learning and Sensory Integration. In J. Cowan, G. Tesauro, J.
Alspector (ed.), Advances in Neural Information Processing Systems 6, 1027-1035. San
Mateo, CA: Morgan Kaufinann.
| 993 |@word trial:1 compression:1 simulation:1 tried:2 dialing:2 reduction:1 initial:1 selecting:1 past:1 current:1 si:1 numerical:3 alone:1 infant:2 half:2 beginning:1 filtered:2 provides:1 wrinkling:1 height:1 profound:1 fps:1 yuhas:2 alspector:1 themselves:1 roughly:2 audiovisual:2 automatically:1 window:1 considering:1 discover:1 moreover:2 underlying:2 emerging:1 developed:1 unified:1 sharpening:1 differentiation:1 temporal:4 control:1 before:1 positive:1 retention:1 local:1 approximately:1 reception:1 twice:1 initialization:1 mateo:1 suggests:1 limited:5 g7:1 practical:1 camera:1 testing:2 lost:2 movellan:5 illusion:1 dodd:1 digit:9 procedure:3 ifthere:1 area:2 significantly:1 composite:2 word:2 integrating:1 seeing:1 convenience:2 s9:1 prentice:1 instability:1 equalization:2 equivalent:1 center:1 starting:3 independently:3 fluent:1 sharpness:1 immediately:1 jackson:2 intermodal:1 variation:1 diego:1 play:2 massaro:4 recognition:24 particularly:1 located:1 role:2 electrical:1 solved:1 initializing:1 compressing:1 revival:1 went:1 asked:3 dynamic:3 hinder:1 trained:5 blurring:2 basis:1 multimodal:1 easily:1 represented:1 america:1 talk:1 jersey:1 describe:1 london:1 sejnowski:2 artificial:4 rho:4 cvpr:1 say:1 compressed:1 favor:1 statistic:2 sequence:2 advantage:2 relevant:4 rapidly:1 holistic:1 poorly:1 representational:3 protrusion:1 impaired:3 regularity:2 cluster:1 enhancement:1 illustrate:1 develop:1 exemplar:3 received:1 indicate:2 meltzoff:2 correct:4 filter:5 stochastic:5 exploration:1 human:13 centered:1 feeding:1 generalization:7 icnn:1 investigation:1 bregler:3 hild:3 hall:1 normal:2 early:1 purpose:1 proc:1 x75:1 create:1 successfully:2 sensor:1 gaussian:14 kaufinann:1 varying:2 focus:1 consistently:1 likelihood:2 centroid:3 entire:4 typically:1 initially:1 hidden:7 pixel:10 overall:2 development:1 integration:3 psychophysics:1 field:1 equal:1 saving:1 discrepancy:1 stimulus:1 simplify:1 few:1 recognize:2 midline:1 individual:2 subsampled:1 phase:1 vowel:2 interest:1 trimming:1 highly:1 evaluation:1 severe:1 male:1 mixture:19 experience:1 old:1 isolated:1 predisposed:1 column:1 disadvantage:1 measuring:2 hearing:7 deviation:4 front:1 dependency:1 synthetic:2 combined:1 density:1 explores:2 sensitivity:1 peak:1 fundamental:1 manke:3 enhance:1 sharpened:1 ear:1 mcgurk:2 cognitive:4 potential:5 student:1 coefficient:1 blurred:3 performed:2 lot:1 portion:1 petajan:1 ir:1 variance:3 who:1 percept:1 characteristic:1 rabiner:2 raw:2 emphasizes:1 ed:2 colleague:1 gain:1 auditory:4 car:2 dimensionality:2 segmentation:1 subtle:1 javier:5 sophisticated:1 goldstein:2 back:1 campbell:1 feed:1 tolerate:1 response:2 improved:2 done:2 evaluated:1 just:2 stage:1 correlation:1 hand:5 mistakenly:1 logistic:3 gray:3 hwang:2 believe:1 pulling:1 effect:8 consisted:2 assigned:2 read:1 laboratory:2 width:1 speaker:4 confusion:5 image:41 physical:1 stork:4 cohen:1 belong:1 approximates:1 numerically:1 refer:2 s5:1 automatic:1 language:1 had:4 stable:1 recent:1 female:1 jolla:1 phone:2 tesauro:1 lipreading:2 morgan:1 captured:1 seen:1 preserving:1 minimum:1 recognized:1 signal:6 sound:1 reduces:1 segmented:1 match:1 long:1 adulthood:1 histogram:2 bimodal:2 achieved:1 cell:2 lea:1 preserved:1 source:1 leaving:2 extra:1 biased:1 subject:15 db:1 cowan:1 constraining:1 vital:1 variety:1 psychology:3 equidistant:1 architecture:6 idea:1 translates:2 unprocessed:2 six:1 effort:4 s7:1 v3a:1 speech:28 processed:1 category:2 reduced:3 percentage:1 nishida:2 s3:1 delta:10 trapped:1 track:4 per:8 estimated:1 four:7 threshold:1 spelke:2 monitor:1 thresholded:1 year:3 powerful:1 audiology:1 reasonable:2 scaling:1 bit:1 capturing:1 followed:1 played:1 display:2 correspondence:1 encountered:1 activity:1 aspect:1 relatively:2 jackknife:1 department:2 according:1 waibel:3 combination:1 poor:1 g3:1 psychologist:1 equation:1 bing:2 montgomery:2 enforcement:3 letting:1 fed:2 end:1 jenkins:2 gaussians:1 kuhl:2 hennecke:4 intelligibility:1 robustness:3 symmetrized:2 voice:1 original:1 clustering:1 cf:6 subsampling:1 macleod:1 build:3 society:1 blend:1 g5:1 macdonald:1 majority:1 acoustical:1 cauchy:3 reason:1 assuming:1 modeled:2 ratio:1 potentially:1 expense:1 sigma:13 perform:1 vertical:1 observation:1 markov:7 variability:2 digitized:3 frame:8 ucsd:2 sentence:2 compensatory:1 california:1 acoustic:10 engine:1 learned:4 adult:1 perception:10 pattern:1 articulation:3 reading:7 hear:1 video:4 event:2 improve:1 movie:2 eye:2 picture:1 axis:1 extract:1 summerfield:2 understanding:2 literature:2 loss:1 rationale:1 filtering:1 age:2 integrate:3 contingency:1 thresholding:4 bank:1 row:1 prone:1 accounted:1 gl:1 last:1 english:4 drastically:1 bias:1 side:1 understand:2 wide:2 face:2 benefit:2 dimension:2 vocabulary:3 calculated:1 transition:1 unaware:1 sensory:2 san:2 preprocessing:11 emphasize:1 continuous:1 table:7 lip:13 nature:1 transfer:1 robust:1 ca:2 symmetry:4 untrained:2 domain:2 did:3 main:2 noise:3 allowed:1 repeated:2 fmd:1 crafted:2 position:1 infancy:1 british:1 decibel:1 supplemented:1 undergraduate:1 modulates:1 illumination:1 explore:1 visual:35 conveniently:1 recommendation:1 underflow:1 conditional:1 month:1 goal:1 kmeans:1 identity:1 room:1 change:1 reducing:1 averaging:1 principal:2 wolff:5 pas:2 tendency:1 la:1 support:1 ongoing:1 audio:4 tested:6 avoiding:1 |
7,236 | 994 | Single Transistor Learning Synapses
Paul Hasler, Chris Diorio, Bradley A. Minch, Carver Mead
California Institute of Technology
Pasadena, CA 91125
(818) 395 - 2812
paul@hobiecat.pcmp.caltech.edu
Abstract
We describe single-transistor silicon synapses that compute, learn,
and provide non-volatile memory retention. The single transistor
synapses simultaneously perform long term weight storage, compute the product of the input and the weight value, and update the
weight value according to a Hebbian or a backpropagation learning
rule. Memory is accomplished via charge storage on polysilicon
floating gates, providing long-term retention without refresh. The
synapses efficiently use the physics of silicon to perform weight updates; the weight value is increased using tunneling and the weight
value decreases using hot electron injection. The small size and
low power operation of single transistor synapses allows the development of dense synaptic arrays. We describe the design, fabrication, characterization, and modeling of an array of single transistor synapses. When the steady state source current is used as
the representation of the weight value, both the incrementing and
decrementing functions are proportional to a power of the source
current. The synaptic array was fabricated in the standard 21'm
double - poly, analog process available from MOSIS.
1
INTRODUCTION
The past few years have produced a number of efforts to design VLSI chips which
"learn from experience." The first step toward this goal is developing a silicon
analog for a synapse. We have successfully developed such a synapse using only
Paul Hasler, Chris Diorio, Bradley A. Minch, Carver Mead
818
Gate
Aoating Gate
!
Drain
Source
...I._High
,oltage
--~~--'-~------r-~--~~--~
Figure 1: Cross section of the single transistor synapse. Our single transistor
synapse uses a separate tunneling voltage terminal The pbase implant results in
a larger threshold voltage, which results in all the electrons reaching the top of the
Si02 barrier to be swept into the floating gate.
a single transistor. A synapse has two functional requirements. First, it must
compute the product of the input multiplied by the strength (the weight) of the
synapse. Second, the synapse must compute the weight update rule. For a Hebbian
synapse, the change in the weight is the time average of the product of the input and
output activity. In many supervised algorithms like backpropagation, this weight
change is the time average of the product of the input and some fed back error
signal. Both of these computations are similar in function. We have developed
single transistor synapses which simultaneously perform long term weight storage,
compute the product of the input and the weight value, and update the weight
value according to a Hebbian or a backpropagation learning rule. The combination
of functions has not previously been achieved with floating gate synapses.
There are five requirements for a learning synapse. First, the weight should be
stored permanently in the absence of learning. Second, the synapse must compute
as an output the product of the input signal with the synaptic weight. Third, each
synapse should require minimal area, resulting in the maximum array size for a given
area. Fourth, each synapse should operate with low power dissipation so that the
synaptic array is not power constrained. And finally, the array should be capable
of implementing either Hebbian or Backpropagation learning rule for modifying
the weight on the floating gate. We have designed, fabricated, characterized, and
modeled an array of single transistor synapses which satisfy these five criteria. We
believe this is the first instance of a single transistor learning synapse fabricated in
a standard process.
2
OVERVIEW
Figure 1 shows the cross section for the single transistor synapse. Since the floating gate is surrounded by Si0 2 , an excellent insulator, charge leakage is negligible
resulting in nearly permanent storage of the weight value. An advantage of using
floating gate devices for learning rules is the timescales required to add and remove
charge from the floating gate are well matched to the learning rates of visual and
auditory signals. In addition, these learning rates can be electronically controlled.
Typical resolution of charge on the floating gate after ten years is four bits (Holler
89). The FETs are in a moderately doped (1 x 10 17 em- 3 ) substrate, to achieve a
Single Transistor Learning Synapses
Vg1
t
819
Vg2
Vtun1
Vd1
(1,1)
t
(1,2)
---+----~--------~----~----VS1~
---+-..-------------+_e_--------- Vtun2
_
Vd2
(2,1)
(2,2)
---+--~~---1---~----VS2~
Figure 2: Circuit diagram of the single - transistor synapse array. Each transistor has a floating gate capacitively coupled to an input column line. A tunneling
connection (arrow) allows weight increase. Weight decreased is achieved by hot
electron injection in the transistor itself. Each synapse is capable of simultaneous
feedforward computations and weight updates. A 2 x 2 section of the array allows
us to characterize how modifying a single floating gate (such as synapse (1,1)) effects the neighboring floating gate values. The synapse currents are a measure of
the synaptic weights, and are summed along each row by the source (Vs) or drain
(Vd) lines into some soma circuit.
high threshold voltage. The moderately doped substrate is formed in the 2pm MOSIS process by the pbase implant. npn transistor. The implant has the additional
benefit of increasing the efficiency of the hot electron injection process by increasing
the electric field in the channel. Each synapse has an additional tunneling junction
for modifying the charge on the floating gate. The tunneling junction is formed
with high quality gate oxide separating a well region from the floating gate.
Each synapse in our synaptic array is a single transistor with its weight stored as
a charge on a floating silicon gate. Figure 2 shows the circuit diagram of a 2 x
2 array of synapses. The column 'gate' inputs (Vg) are connected to second level
polysilicon which capacitively couples to the floating gate. The inputs are shared
along a column. The source (Vs), drain (lid), and tunneling (Viun) terminals are
shared along a row. These terminals are involved with computing the output current
and feeding back 'error' signal voltages. Many other synapses use floating gates to
store the weight value, as in (Holler 89), but none of the earlier approaches update
the charge on the floating gate during the multiplication of the input and floating
gate value. In these previous approaches one must drive the floating gate over large
a voltage range to tunnel electrons onto the floating gate. Synaptic computation
must stop for this type of weight update.
The synapse computes as an output current a product of weight and input signal,
820
Paul Hasler. Chris Diorio. Bradley A. Minch. Carver Mead
10.7
Tunneling Opel1llions
synapse (1.2)
... . . . '.
Iojec:tion Opel1lliolll
10"
synapse (2,1), synapse (2.2)
10"
". "" synapse (1.1)
... ..-
15
S
j
10. 10
10.11
10- 12
row 2 background alln:nl
row I IJeckarouod alrlent
10-13 "-_--'-_ _-'-_ _-'--_ _'"--_----'_ _--'-_ _...J
o
SO
I 00
I SO
200
2S0
300
3SO
Step iteration Number
Figure 3: Output currents from a 2 x 2 section of the synapse array, showing
180 injection operations followed by 160 tunneling operations. For the injection
operations, the drain (V dl) is pulsed from 2.0 V upto 3.3 V for 0.5s with Vg1 at 8V
and Vg2 at OV. For the tunneling operations, the tunneling line (Vtunl) is pulsed
from 20 V up to 33.5 V with Vg2 at OV and "91 at 8V. Because our measurements
from the 2 x 2 section come from a larger array, we also display the 'background'
current from all other synapses on the row. This background current is several orders
of magnitude smaller than the selected synapse current, and therefore negligible.
and can simultaneously increment or decrement the weight as a function of its input
and error voltages. The particular learning algorithm depends on the circuitry at
the boundaries of the array; in particular the circuitry connected to each of the
source, drain, and tunneling lines in a row. With charge Qlg on the floating gate
and Vs equal to 0 the subthreshold source current is described by
(1)
k;.
where Qo is a device dependent parameter, and UT is the thermal voltage
The coupling coefficient, 6, of the gate input to the transistor surface potential is
typically less than 0.1. From ( 1) We can consider the weight as a current I, defined
by
ISllnapse
=
(
~~)
Ioe---qo- e
T
e
6gVp
T
= Ie
6gVp
T
(2)
where Vgo is the input voltage bias, and ~ Vg is Vg - "90. The synaptic current is
thus the product of the weight, I, and a weak exponential function of the input
voltage.
The single transistor learning synapses use a combination of electron tunneling and
hot electron injection to adapt the charge on the floating gate, and thereby the
821
Single Transistor Learning Synapses
weight of the synapse. Hot electron injection adds electrons to the floating gate,
thereby decreasing the weight. Injection occurs for large drain voltages; therefore
the floating gate charge can be reduced during normal feedforward operation by raising the drain voltage. Electron tunneling removes electrons from the floating gate,
thereby increasing the weight. The tunneling line controls the tunneling current;
thus the floating gate charge can be increased during normal feedforward operation
by raising the tunneling line voltage. The tunneling rate is modulated by both the
input voltage and the charge on the floating gate.
Figure 3 shows an example the nature of the weight update process. The source
current is used as a measure of the synapse weight. The experiment starts with all
four synapses set to the same weight current. Then, synapse (1,1) is injected for 180
cycles to preferentially decrease its weight. Finally, synapse (1,1) is tunneled for 160
cycles to preferentially increase its weight. This experiment shows that a synapse
can be incremented by applying a high voltage on tunneling terminals and a low
voltage on the input, and can be decremented by applying a high voltage on drain
terminals and a high voltage on the input. In the next two sections, we consider
the nature of these update functions. In section three we examine the dependence
of hot electron injection on the source current of our synapses. In section four we
examine the dependence of electron tunneling on the source current of our synapses.
3
Hot Electron Injection
Hot electron injection gives us a method to add electrons to the floating gate. The
underlying physics of the injection process is to give some electrons enough energy
and direction in the channel to drain depletion region to surmount the Si0 2 energy
barrier. A device must satisfy two requirements to inject an electron on a floating
gate. First, we need a region where the potential drops more than 3.1 volts in a
distance of less than 0.2pm to allow electrons to gain enough energy to surmount
the oxide barrier. Second, we need a field in the oxide in the proper direction
to collect electrons after they cross the barrier. The moderate substrate doping
level allows us to easily achieve both effects in subthreshold operation. First, the
higher substrate doping results in a much higher threshold voltage (6.1 V), which
guarantees that the field in the oxide at the drain edge of the channel will be in
the proper direction for collecting electrons over the useful range of drain voltages.
Second, the higher substrate doping results in higher electric fields which yield
higher injection efficiencies. The higher injection efficiencies allow the device to have
a wide range of drain voltages substantially below the threshold voltage. Figure 4
shows measured data on the change in source current during injection vs. source
current for several values of drain voltage.
Because the source current, I, is related to the floating gate charge, Q,g as shown
in ( 1) and the charge on the floating gate is related to the tunneling or injection
current (I, g) by
dQ,g _ I
(3)
dt - Ig
an approximate model for the change of the weight current value is
I
(4)
-dl
=
-1
dt
Qo
'9
Paul Hasler, Chris Diorio, Bradley A. Minch, Carver Mead
822
!
10"
i
10.10
~
10'"
synapse (1,1)
i!,
J.
.S
3.3
10.12
g
synapse (1,2)
10'"
3.6
Sour.-e Curren! (Weiaht Value)
Figure 4: Source Current Decrement during injection vs. Source Current for several
values of drain voltage. The injection operation decreases the synaptic weight. V 92
was held at OV, and V 91 was at 8V during the 0.5s injecting pulses. The change in
source current is approximately proportional to the source current to the f3 power,
where of f3 is between 1.7 and 1.85 for the range of drain voltages shown. The change
in source current in synapse (1,2) is much less than the corresponding change in
synapse (1,1) and is nearly independent of drain voltage. The effect of this injection
on synapses (2,1) and (2,2) is negligible.
The injection current can be approximated over the range of drain voltages shown
in Fig. 4 by (Hasler 95)
(5)
where Vd-c is the voltage from the drain to the drain edge of the channel, Vd is the
drain voltage, /0 is a slowly varying function defined in (Hasler 95), and Vini is in
the range of 60m V to 100mV. A is device dependent parameter, Since hot electron
injection adds electrons to the floating gate, the current into the floating gate (If 9)
is negative, which results in
IfJ ~
-dI
=
-A-e .nJ
(6)
dt
Qo
The model agrees well with the data in Fig. 4, with f3 in the range of 1. 7 - 1.9.
Injection is very selective along a row with a selectivity coefficient between 102 and
107 depending upon drain voltage and weight. The injection operations resulted in
negligible changes in source current for synapses (2,1) and (2,2).
4
ELECTRON TUNNELING
Electron tunneling gives us a method for removing electrons from the floating gate.
Tunneling arises from the fact that an electron wavefunction has finite extent. For a
Single Transistor Learning Synapses
i
10"
I
'Ii
i!
10.9
~
VlUn
I
a
)10.
823
=36.0 ?'
3S.0 ..--
10
34.0
3S.S
34.S
.5
110.
11
33.S
33.0
Source Cumnt (Weig/ll Value)
Figure 5: Synapse (1,1) source current increment vs. Source Current for several
values of tunneling voltage. The tunneling operation increases the synaptic weight.
V 91 was held at OV and V 92 was 8V while the tunneling line was pulsed for 0.5s
from 20V to the voltage shown. The change in source current is approximately
proportional to the a power of the svurce current where a is between 0.7 and 0.9
for the range of tunneling voltages shown. The effect of this tunneling procedure
on synapse (2,1) and (2,2) are negligible. The selectivity ratio of synapses on the
same row is typically between 3-7 for our devices.
thin enough barrier, this extent is sufficient for an electron to penetrate the barrier.
An electric field across the oxide will result in a thinner barrier to the electrons on
the floating gate. For a high enough electric field, the electrons can tunnel through
the oxide.
When traveling through the oxide, some electrons get trapped in the oxide, which
changes the barrier profile. To reduce this trapping effect we tunnel through high
quality gate oxide, which has far less trapping than interpoly oxide. Both injection
and tunneling have very stable and repeatable characteristics. When tunneling at
a fixed oxide voltage, the tunneling current decreases only 50 percent after lOnG of
charge has passed through the oxide. This quantity of charge is orders of magnitude
more than we would expect a synapse to experience over a lifetime of operation.
Figure 5 shows measured data on the change in source current during tunneling as
a function of source current for several values of tunneling voltage. The functional
form of tunneling current is of the form (Lenzlinger 69)
(7)
where Yo, f Olun are model parameters which roughly correspond with theory. Tunneling removes electrons from the floating gate; therefore the floating gate current
is positive. By expanding Vfg for fixed Viun as VfgO + ~ Vfg and inserting ( 1), the
824
Paul Hasler, Chris Diorio, Bradley A. Minch, Carver Mead
II Parameter I Typical Values II Parameter I Typical Values II
{3
Q'
1.7 - 1.9
0.7 - 0.9
.2 pC
6
0.02 - 0.1
Qo
8.6 x 10 ?~u
A
78mV
Vinj
Table 1: Typical measured values of the parameters in the modeling of the single
transistor synapse array.
resulting current change is
dI
I
dt =
Q:n e- Viu
Vg
v/ gO Iso
(
I
Iso
)Q
(8)
where IsO is the bias current corresponding to VfgO. The model qualitatively agrees
with the data in Fig. 5, with Q' in the range of 0.7- 0.9. The tunneling selectivity
between synapses on different rows is very good, but tunneling selectivity along
along a row is poor. We typically measure tunneling selectivity ratios along a row
between 3 - 7 for our devices.
5
Model of the Array of Single Transistor Synapses
Finally, we present an approximate model of our array of these single transistor
synapses. The learning increment of the synapse at position (i,i) can be modeled
as
61l.V
ISllnapse'j = Iije ~ == Iso WijXj
dW. .
1
_
Vq
_
;d;
_
(9)
=..:.:...!.L =.:.?t.II..A.e VI"n v/ gO W~x~ 1 _....d.... e inj WP.xl? 1
dt
Qo
I) )
Qo
I) )
for the synapse at position (i,i), where Wi,j can be considered the weight value,
and x j are the effective inputs network. Typical values for the parameters in ( 9 )
are given in Table 1.
Acknowledgments
The work was supported by the office of Naval Research, the Advanced Research
Projects Agency, and the Beckman Foundation.
References
P. Hasler, C. Diorio, B. Minch, and C. Mead (1995) "An Analytic model of Hot
Electron Injection from Boltzman Transport", Tech. Report 123456
M. Holler, S. Tam, H. Castro, and R. Benson (1989), "An electrically trainable
artificial neural network with 10240 'floating gate' synapses", International Joint
Conference on Neural Networks, Wasllington, D.C., June 1989, pp. 11-191 - 11-196.
M. Lenzlinger and E. H. Snow (1969), "Fowler-Nordheim tunneling into thermally
grown Si0 2 ," J. Appl. Phys., vol. 40, pp. 278-283, 1969.
PARTVll
SPEECH AND SIGNAL PROCESSING
| 994 |@word pulse:1 thereby:3 past:1 vg2:3 bradley:5 current:41 must:6 refresh:1 analytic:1 remove:3 designed:1 drop:1 update:9 v:6 selected:1 device:7 trapping:2 iso:4 characterization:1 five:2 along:7 roughly:1 examine:2 terminal:5 decreasing:1 vfg:2 increasing:3 project:1 matched:1 underlying:1 circuit:3 substantially:1 developed:2 gvp:2 fabricated:3 sour:1 nj:1 guarantee:1 collecting:1 charge:16 control:1 positive:1 retention:2 negligible:5 thinner:1 mead:6 approximately:2 collect:1 appl:1 range:9 acknowledgment:1 backpropagation:4 procedure:1 area:2 get:1 onto:1 storage:4 applying:2 go:2 resolution:1 penetrate:1 rule:5 ioe:1 array:17 pbase:2 dw:1 increment:3 substrate:5 us:1 approximated:1 vd2:1 region:3 connected:2 cycle:2 diorio:6 decrease:4 incremented:1 agency:1 moderately:2 ov:4 upon:1 efficiency:3 easily:1 joint:1 chip:1 grown:1 describe:2 effective:1 artificial:1 larger:2 itself:1 advantage:1 transistor:25 product:8 neighboring:1 inserting:1 achieve:2 doped:2 double:1 requirement:3 vinj:1 coupling:1 depending:1 measured:3 come:1 vlun:1 direction:3 snow:1 modifying:3 vfgo:2 implementing:1 require:1 feeding:1 considered:1 normal:2 electron:33 circuitry:2 injecting:1 beckman:1 si0:3 agrees:2 successfully:1 reaching:1 varying:1 voltage:34 office:1 yo:1 june:1 naval:1 tech:1 dependent:2 typically:3 pasadena:1 vlsi:1 selective:1 development:1 constrained:1 summed:1 field:6 equal:1 f3:3 nearly:2 thin:1 report:1 decremented:1 few:1 simultaneously:3 resulted:1 floating:38 nl:1 pc:1 held:2 edge:2 capable:2 experience:2 carver:5 capacitively:2 minimal:1 increased:2 instance:1 modeling:2 column:3 earlier:1 doping:3 fabrication:1 characterize:1 stored:2 minch:6 international:1 ie:1 physic:2 e_:1 holler:3 slowly:1 oxide:12 inject:1 tam:1 potential:2 coefficient:2 satisfy:2 permanent:1 mv:2 depends:1 vi:1 tion:1 vs2:1 start:1 formed:2 characteristic:1 efficiently:1 subthreshold:2 yield:1 correspond:1 weak:1 produced:1 none:1 drive:1 simultaneous:1 synapsis:26 phys:1 synaptic:10 energy:3 pp:2 involved:1 vgo:1 di:2 couple:1 stop:1 auditory:1 gain:1 lenzlinger:2 ut:1 back:2 higher:6 dt:5 supervised:1 synapse:42 lifetime:1 traveling:1 qo:7 transport:1 thermally:1 quality:2 fowler:1 believe:1 effect:5 volt:1 wp:1 ll:1 during:7 steady:1 fet:1 criterion:1 dissipation:1 percent:1 volatile:1 functional:2 overview:1 analog:2 silicon:4 measurement:1 pm:2 stable:1 surface:1 add:4 pulsed:3 moderate:1 store:1 selectivity:5 accomplished:1 swept:1 caltech:1 additional:2 wijxj:1 signal:6 ii:5 hebbian:4 polysilicon:2 characterized:1 adapt:1 long:4 cross:3 controlled:1 iteration:1 achieved:2 addition:1 background:3 decreased:1 diagram:2 source:25 operate:1 feedforward:3 enough:4 npn:1 curren:1 reduce:1 passed:1 effort:1 vd1:1 speech:1 tunnel:3 useful:1 ten:1 reduced:1 trapped:1 vol:1 four:3 soma:1 threshold:4 ifj:1 hasler:8 mosis:2 year:2 surmount:2 fourth:1 injected:1 tunneling:39 bit:1 followed:1 display:1 activity:1 strength:1 injection:25 developing:1 according:2 combination:2 poor:1 electrically:1 smaller:1 across:1 em:1 wi:1 lid:1 castro:1 benson:1 depletion:1 vq:1 previously:1 vs1:1 fed:1 junction:2 operation:12 available:1 multiplied:1 upto:1 gate:44 permanently:1 top:1 pcmp:1 leakage:1 quantity:1 occurs:1 dependence:2 distance:1 separate:1 separating:1 vd:3 chris:5 extent:2 toward:1 modeled:2 providing:1 ratio:2 preferentially:2 negative:1 design:2 proper:2 perform:3 finite:1 thermal:1 required:1 connection:1 raising:2 california:1 viu:1 nordheim:1 below:1 memory:2 hot:10 power:6 advanced:1 technology:1 coupled:1 drain:21 multiplication:1 wavefunction:1 expect:1 proportional:3 vg:4 foundation:1 sufficient:1 s0:1 dq:1 surrounded:1 row:11 supported:1 electronically:1 bias:2 allow:2 institute:1 wide:1 barrier:8 benefit:1 boundary:1 computes:1 qualitatively:1 ig:1 far:1 boltzman:1 approximate:2 decrementing:1 table:2 learn:2 channel:4 nature:2 ca:1 expanding:1 excellent:1 poly:1 electric:4 dense:1 timescales:1 arrow:1 incrementing:1 decrement:2 paul:6 profile:1 fig:3 position:2 exponential:1 xl:1 third:1 removing:1 repeatable:1 showing:1 dl:2 magnitude:2 implant:3 visual:1 goal:1 shared:2 absence:1 change:12 typical:5 inj:1 modulated:1 arises:1 trainable:1 |
7,237 | 996 | Bias, Variance and the Combination of
Least Squares Estimators
Ronny Meir
Faculty of Electrical Engineering
Technion, Haifa 32000
Israel
rmeirGee.technion.ac.il
Abstract
We consider the effect of combining several least squares estimators
on the expected performance of a regression problem. Computing
the exact bias and variance curves as a function of the sample size
we are able to quantitatively compare the effect of the combination
on the bias and variance separately, and thus on the expected error
which is the sum of the two. Our exact calculations, demonstrate
that the combination of estimators is particularly useful in the case
where the data set is small and noisy and the function to be learned
is unrealizable. For large data sets the single estimator produces
superior results. Finally, we show that by splitting the data set
into several independent parts and training each estimator on a
different subset, the performance can in some cases be significantly
improved.
Key words: Bias, Variance, Least Squares, Combination.
1
INTRODUCTION
Many of the problems related to supervised learning can be boiled down to the
question of balancing bias and variance. While reducing bias can usually be accomplished quite easily by simply increasing the complexity of the class of models
studied, this usually comes at the expense of increasing the variance in such a way
that the overall expected error (which is the sum of the two) is often increased.
296
Ronny Meir
Thus, many efforts have been devoted to the issue of decreasing variance, while attempting to keep the concomitant increase in bias as small as possible. One of the
methods which has become popular recently in the neural network community is
variance reduction by combining estimators, although the idea has been around in
the statistics and econometrics literature at least since the late sixties (see Granger
1989 for a review). Nevertheless, it seems that not much analytic work has been
devoted to a detailed study of the effect of noise and an effectively finite sample size
on the bias/variance balance. It is the explicit goal of this paper to study in detail
a simple problem of linear regression, where the full bias/variance curve can be
computed exactly for any effectively finite sample size and noise level. We believe
that this simple and exactly solvable model can afford us insight into more complex
non-linear problems, which are at the heart of much of the recent work in neural
networks.
A further aspect of our work is related to the question of the partitioning of the
data set between the various estimators. Thus, while most studies assume the each
estimator is trained on the complete data set, it is possible to envisage a situation
where the data set is broken up into several subsets, using each subset of data to form
a different estimator. While such a scheme seems wasteful from the bias point of
view, we will see that in fact it produces superior foreca..,ts in some situations. This,
perhaps suprising, result is due to a large decrease in variance resulting from the
independence of the estimators, in the case where the data subsets are independent.
2
ON THE COMBINATION OF ESTIMATORS
The basic objective of regression is the following: given a finite training set, D, composed of n input/output pairs, D = {(XIJ' YIJ)}~=l' drawn according to an unkown
distribution P(x, y), find a function (,estimator'), f(x; D), which 'best' approximates y. Using the popular mean-squared error criterion and taking expectations
with respect to the data distribution one finds the well-known separation of the
error into a bias and variance terms respectively (Geman et al. 1992)
+ ED (f(x; D) - EDf(x; D)]2 .
(1)
We consider a data source of the form y = g(x) + 7], where the 'target' function g(x)
?(x) = (EDf(x; D) - E[ylx])2
is an unknown (and potentially non-linear) function and 7] is a Gaussian random
variable with zero mean and variance (72. Clearly then E[Ylx] = g(x).
In the usual scenario for parameter estimation one uses the complete data set, D,
to form an estimator f(x; D). In this paper we consider the case where the data set
D is broken up into I< subsets (not necessarily disjoint), such that D =
1 DC k) ,
and a separate estimator is found for each subset. The full estimator is then given
by the linear combination (Granger 1989)
nf=
K
f(x; D) =
L bk/k(x; D(k)
(2)
k=l
The optimal values of the parameters bk can be easily obtained if the data distribution, P(x, y), is known, by simply minimizing the mean-squared error (Granger
Bias. Variance and the Combination of Least Squares Estimators
297
1989). In the more typical case where this distribution is unkown, one may
resort to other schemes such as least-squares fitting for the parameter vector
b = {b l , ... , bK}. The bias and variance of the combined estimator can be simply
expressed in this case, and are given by
B(x; g)
=
(t
k=l
hh:(x) - g(x))
2
;
V(x; g)
= L bkh ' {lkfk'(X) -
fk(X)fk'(X)}
k,k'
(3)
where the overbars denote an average with respect to the data. It is immediately
apparent that the variance term is composed of two contributions. The first term,
corresponding to k = k', simply computes a weighted average of the single estimator
variances, while the second term measures the average covariance between the different estimators. While the first term in the variance can be seen to decay as 1/ J{
in the case where all the weights h are of the same order of magnitude, the second
term is finite unless the covariances between estimators are very small. It would
thus seem beneficial to attempt to make the estimators as weakly correlated as possible in order to decrease the variance. Observe that in the extreme case where the
data sets are independent of each other, the second term in the variance vanishes
identically. Note that the bias term depends only on single estimator properties
and can thus be computed from the theory of the single estimator. As mentioned
above, however , the second term in the variance expression explicitly depends on
correlations between the different estimators, and thus requires the computation of
quantities beyond those of single estimators.
3
THE SINGLE LINEAR ESTIMATOR
Before considering the case of a combination of estimators, we first review the case
of a single linear estimator, given by f(x; D) = wT . x, where w is estimated from
the data set D. Following Bos et al. (1993) we further assume that the data arises
through an equation of the form y = g(x) + 1] with 9 = g(w~ . x). Looking back
at equations (3) it is clear that the bias and variance are explicit functions of x
and the weight vector woo In order to remove the explicit dependence we compute
in what follows expectations with respect to the probability distribution of x and
Wo, denoted respectively by E p ['] and E o[']' Thus, we define the averaged bias and
variance by B = EoEp[B(x; wo)] and V = EoEp[V(x; wo)] and the expected error
is then ? = B + V.
In this work we consider least-squares estimation which corresponds to minimizing
the empirical error, ?emp(w, D) = IIXw - Y112, where X is the n x d data matrix,
Y is the n x 1 output vector and w is a d x 1 weight vector. The components of
the 'target' vector Yare given by YP = g(w~ . xl-') + 1]p where 1Jp are i.i.d normal
random variables with zero mean and variance (J'2 . Note that while we take the
estimator itself to be linear we allow the target function g(.) to be non-linear. This
is meant to model the common situation where the model we are trying to fit is
inadequate, since the correct model (even it exists) is usually unkown.
Thus, the least squares estimator is given by wE argminw?emp(w,D). Since in
this case the error-function is quadratic it possesses either a unique global minimum
298
Ronny Meir
or a degenerate manifold of minima, in the case where the Hessian matrix, XT X,
is singular.
The solution to the least squares problem is well known (see for example Scharf
1991), and will be briefly summarized. When the number of examples, n, is smaller
than the input dimension, d, the problem is underdetermined and there are many
solutions with zero empirical error. The solutions can be written out explicitly in
the form
(n < d),
(4)
where V is an arbitrary d-dimensional vector. It should be noted that any vector w
satisfying this equation (and thus any least-squares estimator) becomes singular as
n approaches d from below, since the matrix X XT becomes singular. The minimal
norm solution, often referred to as the Moore-Penrose solution, is given in this case
by the choice V
O. It is common ill the literature to neglect the study of the
underdetermined regime since the solution is not unique in this case. We however
will pay specific attention to this case, corresponding to the often prevalent situation where the amount of data is small, attempting to show that the combination
of estimators approach can significantly improve the quality of predictors in this
regime. Moreover, many important inverse problems in signal processing fall into
this category (Scharf 1991).
=
In the overdetermined case, n > d (assuming the matrix X to be of full rank),
a zero error solution is possible only if the function g( .) is linear and there is no
noise, namely El1]2] = O. In any other case, the problem becomes unrealizable and
the minimum error is non-zero. In any event, in this regime the unique solution
minimizing the empirical error is given by
W = (XT X)-l XTy
(n > d).
(5)
It is eay to see that this estimator is unbiased for linear g( .).
In order to compute the bias and variance for this model we use Eqs. (3) with
K
1 and bTl:
1. In order to actually compute the expectations with respect to
x and the weight vector Wo we assume in what follows that the random vector x
is distributed according to a multi-dimensional normal distributions of zero mean
and covariance matrix (lid)!. The vector Wo is similarly distributed with unit
covariance matrix. The reason for the particular scaling chosen for the covariance
matrices will become clear below. In the remainder of the paper we will be concerned
with exact calculations in the so called thermodynamic limit: n, d -+ 00 and a =
nl d finite. This limit is particularly useful in that the central limit theorem allows
one to make precise statements about the behavior of the system, for an effectively
finite sample size, a. We note in passing that in the thermodynamic limit, d -+ 00,
we have Ei
-+ 1 with probability 1 and similarly for (lid) Ei W5i' Using these
simple distributions we can, after some algerbra, directly compute the bias and
variance. Denoting R = Eo[w T . wo], r = Eollwll 2 , Q = Eollwl1 2 , one can show
that the bias and variance are given by
=
=
x;
B
=
7' -
2ugR + g2
v = Q -r.
(6)
=
In the above equations we have used g2 = f Dug 2(u) and ug f Du ug(u) where
the Gaussian measure Du is defined by Du = (e-u~/2 1...j2-i)du. We note in passing
Bias, Variance and the Combination of Least Squares Estimators
299
that the same result is obtained for any i.i.d variables, Xi, with zerO mean and
variance lid. We thus note that a complete calculation of the expected bias and
variance requires the explicit computation of the variables R, rand Q defined
above. In principle, with the explicit expressions (4) and (5) at hand one may
proceed to compute all the quantities relevant to the evaluation of the bias and
variance. Unfortunately, it turns out that a direct computation of r, Rand Q using
these expressions is a rather difficult task in the theory of random matrices, keeping
in mind the potential non-linearity of the function g{.). A way to solve the problem
can be undertaken via a slightly indirect route, using tools from statistical physics.
The variables Rand Q above have been recently computed by Bos et al. (1993)
using replicas and by Opper and Kinzel (1994) by a direct calculation. The variable
r can be computed along similar lines resulting in the following expressions for the
bias and variance (given for brevity for the Moore-Penrose solution):
0'< 1 :
v=
B
_0'_ [g2
1-0'
+ u2 _
a> 1 :
0'(2 _ a )ttg2]
(7)
We see from this solution that for Q > 1 the bias is constant, while the variance
is monotonically decreasing with the sample size Q. For Q < 1, there are of course
multiple solutions corresponding to different normalizations Q. It is easy to see,
however, that the Moore-Penrose solution, gives rise to the smallest variance of all
least-squares estimators (the bias is unaffected by the normalization of the solution) .
The expected (or generalization) error is given simply by ? = B + V, and is thus
smallest for the Moore-Penrose solution. Note that this latter result is independent
of whether the function g(.) is linear or not. We note in passing that in the simple
case where the target function g(-) is linear and the data is noise-free (u 2 = 0) one
obtains the particularly simple result ? = 1 - a for a < 1 and ? = 0 above Q = 1.
Note however that in any other case the expected error is a non-linear function of
the normalized sample size a.
4
COMBINING LINEAR ESTIMATORS
Having summarized the situation in the case of a single estimator, we proceed to
the case of K linear estimators. In this case we assume that the complete data
set, D, is broken up into K subsets D(k) of size nk = Qkd each. In particular we
consider two extreme situations: (i) The data sets are independent of each other ,
namely D(k) n D(k') = 0 for all k, and (ii) D(k) = D for all k. We refer to the first
situation as the non-overlapping case, while to the second case where all estimators
are trained on the same data set as the fully-overlappin~ case. Denoting by w(k)
the k'th least-squares estimator, based on training set D( ), we define the following
quantities:
R(k)
= Eo [w(k)T . w o]
, r(k,k')
= Eo [w(k)T . w(k')]
, p(k,k')
= Eo [w(k)T . w(k')] .
(8)
Making use of Eqs. (3) and the probability distribution of Wo one then straighfor-
300
Ronny Meir
wardly finds for the mixture estimator
L
L b~
blebk,r(k,Ie') - 2ug
1e,Ie'
L
bkR(k)
+ g2,
Ie
[p<k,le) - r(k,Ie)]
Ie
+ L:
bleb le , [p(k,k') _ r(k,k')] .
(9)
k#Ie'
Now, the computation of the parameters p(le,k) and R(k) is identical to the case of
a single estimator and can be directly read from the results of Bos et al. (1993)
with the single modification of rescaling the sample size a to ak. The only interaction between estimators arises from the terms p(Ie,k') and r(k ,k') which couple
two different estimators. Note however that r(k,k') is independent of the degree of
overlap between the data sets D(Ie) since each estimator is averaged independently.
In fact , a straighforward, if lengthy, calculation leads to the conclusion that in general r(Ie,Ie') = R(Ie) R(1e'). For the case where the data sets are independent of each
other, namely D(k) n D(k') = 0 it is obvious that p(k ,k')
r(k,k'). One then finds
p(Ie,Ie') = r(Ie,Ie') = R(Ie) R(k') for independent data sets. In the case where the data
sets are identical, D(k) = D for all k, the variable p(k,k') has been computed by
BOs et al. (1993). At this point we proceed by discussing individually the two data
generation scenarios described in section 1.
=
4.1
NON?OVERLAPPING DATA SETS
As shown above, since the data sets are independent of each other one has in this
case p(le,k') = r(k,k'), implying that the second term in the expression for the variance
in Eq. (9) vanishes. Specializing now to the case where the sizes of the data sets
are equal, namelyak
nk/d
a/I{ for all k, we expect the variables R(k) and
O(k) to be independent of the specific index k. In the particular case of a uniform
mixture (b k = 1/1< for all k) we obtain
=
=
(10)
where the subscript 1< indicates that the values of the corresponding variables must
be taken with respect to aK = a/1<. The specific values of the parameters RK
and OK can be found directly from the solution of Bos et al. (1993) by replacing
a by a/I{ and dividing the variance term by a further ]{ factor. Since the bias
is a non-increasing function a it is always increased (or at best unchanged) by
combining estimators from non-overlapping data subsets. It can also be seen from
the above equation that as 1< increases the variance can be made arbitrarily small,
albeit at the expense of increasing the bias. It can thus be expected that for any
sample size a there is an optimal number of estimators minimizing the expected
error, ? = B + V. In fact, for small a one may expand the equations for the bias
and variance obtaining the expected error from which it is easy to see that the
optimal value of 1< is given by 1<* = (g2 + (72)/ug 2. As could be expected we
see that the optimal value of I{ scales with the noise level (72, since the effect of
combining multiple independent estimators is to reduce the effective noise variance
by a factor of I{. As the sample size a increases and outweighs the effect of the
noise, we find that the bias increase due to the data partitioning dominates the
Bias, Variance and the Combination of Least Squares Estimators
301
decrease in variance and the combined estimator performs more and more poorly
relative to the single estimator. Finally, for a > K we find the choice 1< = 1 always
yields a lower expected error. Thus, while we have shown that for small sample
sizes the effect of splitting the data set into independent subsets is helpful, this
is no longer the case if the sample size is sufficiently large, in which case a single
estimator based on the complete data set is superior. For a -+ 00, however , one
finds that all uniform mixtures converge (to leading order) at the same rate, namely
?K(a) : : : : ?00 + (?00 + (!2)ja , where ?00 = g2 - ug 2. For finite values of cr, however,
the value of 1< has a strong effect on the quality of the stimator.
4.2
FULLY OVERLAPPING DATA SETS
We focus now on the case where all estimators are formed from the same data set
D, namely DUe) = D for all k . Since there is a unique solution to the least-squares
estimation problem for a > 1, all least-squares estimators coincide in this regime .
Thus, we focus here on the case a < 1, where multiple least-squares estimators
coexist. We further assume that only mixtures of estimators of the same norm Q
are allowed. We obtain for the uniform mixture
Clearly the expression for the bias in this case is identical to that obtained for
the single estimator, since all estimators are based On the same data set and the
bias term depends only on single estimator properties. The variance term , however ,
is modified due to the correlation between the estimators expressed through the
variable p(k,k'). Since the variance for the case of a single estimator is Q - R2 and
since q $ Q it is clear in this case that the variance is reduced while the bias remains
unchanged. Thus we conclude that the mixture of estimators in this case indeed
produces superior performance to that of the single estimator. However, it can be
seen that in the case of the Moore-Penrose solution , corresponding to choosing the
smallest possible norm Q, the expected error is minimal. We thus conclude that for
a < 1 the Moore-Penrose pseudo-inverse solution yields the lowest expected error ,
and this cannot be improved on by combining least-squares estimators obtained
from the full data set D .
Recall that we have shown in the previous section that (for small and noisy data
sets) combining estimators formed using non-overlapping data subsets produced
results superior to those of any single estimator trained on the complete data set.
An interesting conclusion of these results is that splitting the data set into nonoverlapping subsets is a better strategy than training each estimator with the full
data. As mentioned previously, the basic reason for this is the independence of
the estimators formed in this fashion, which helps to reduce the variance term
more drastically than in the case where the estimators are dependent (having been
exposed to overlapping data sets).
5
CONCLUSIONS
In this paper we have studied the effect of combining different estimators on the
performance of linear regression. In particular we have focused on the case of linear
302
Ronny Meir
least-squares estimation, computing exactly the full bias and variance curves for the
case where the input dimension is very large (the so called thermodynamic limit).
While we have focused specifically on the case of linear estimators, it should not be
hard to extend these results to simple non-linear functions ofthe form f(w T ?x) (see
section 2). The case of a combination of more complex estimators (such as multilayered neural networks) is much more demanding, as even the case of a single such
network is rather difficult to analyzes.
Several positive conclusions we can draw from our study are the following. First, the
general claim that combining experts is always helpful is clearly fallacious. While
we have shown that combining estimators is beneficial in some cases (such as small
noisy data sets), this is not the case in general. Second, we have shown that in some
situations (specifically unrealizable rules and small sample size) it is advantageous
to split the data into several non-overlapping subsets. It turns out that in this case
the decrease in variance resulting from the independence of the different estimators,
is larger than the concomitant increase in bias. It would be interesting to try to
generalize our results to the case where the data is split in a more efficient manner.
Third, our results agree with the general notion that when attempting to learn
an unrealizable function (whether due to noise or to a mismatch with the target
function) the best option is to learn with errors.
Ultimately one would like to have a general theory for combining empirical estimators. Our work has shown that the effect of noise and finite sample size is expected
to produce non-trivial effects which are impossible to observe when considering only
the asymptotic limit.
Acknowledgements
The author thanks Manfred Opper for a very helpful conversation and the Ollendorff center of the Electrical Engineering department at the Technion for financial
support.
References
S. Bos., W. Kinzel and M. Opper 1993, The generalization ability of perceptrons
with continuous outputs, Phys. Rev. A 47:1384-1391.
S. Geman, E. Bienenstock and R. Dorsat 1992, Neural networks and the
bias/variance dilemma, Neural Computation 4:1-58.
C.W.J. Granger 1989, Combining forecasts - twenty years later, J. of Forecast.
8:167-173.
M. Opper and W. Kinzel 1994, Statistical mechanics of generalization, in Physics
of Neural networks, van Hemmen, J.S., E. Domany and K. Schulten eds., SpringerVerlag, Berlin.
L.1. Scharf, 1991 Statistical Signal Processing: Detection, Estimation and Time
Series Analysis, Addison-Wesley, Massachusetts.
| 996 |@word briefly:1 faculty:1 advantageous:1 seems:2 norm:3 covariance:5 reduction:1 series:1 denoting:2 written:1 must:1 analytic:1 remove:1 implying:1 el1:1 manfred:1 along:1 direct:2 become:2 fitting:1 manner:1 indeed:1 expected:15 behavior:1 mechanic:1 multi:1 decreasing:2 considering:2 increasing:4 becomes:3 moreover:1 linearity:1 lowest:1 israel:1 what:2 pseudo:1 nf:1 exactly:3 partitioning:2 unit:1 before:1 positive:1 engineering:2 limit:6 ak:2 subscript:1 studied:2 averaged:2 unique:4 empirical:4 significantly:2 word:1 cannot:1 coexist:1 ronny:5 impossible:1 center:1 attention:1 independently:1 focused:2 splitting:3 immediately:1 estimator:75 insight:1 rule:1 financial:1 notion:1 target:5 exact:3 us:1 overdetermined:1 satisfying:1 particularly:3 econometrics:1 geman:2 electrical:2 decrease:4 mentioned:2 vanishes:2 broken:3 complexity:1 ultimately:1 trained:3 weakly:1 exposed:1 dilemma:1 easily:2 indirect:1 various:1 effective:1 choosing:1 quite:1 apparent:1 larger:1 solve:1 ability:1 statistic:1 noisy:3 envisage:1 itself:1 unkown:3 interaction:1 remainder:1 argminw:1 j2:1 relevant:1 combining:12 degenerate:1 poorly:1 produce:4 help:1 ac:1 eq:3 strong:1 dividing:1 suprising:1 come:1 correct:1 ja:1 generalization:3 bkr:1 unrealizable:4 underdetermined:2 yij:1 around:1 sufficiently:1 normal:2 claim:1 smallest:3 estimation:5 individually:1 tool:1 weighted:1 clearly:3 gaussian:2 always:3 modified:1 rather:2 cr:1 focus:2 prevalent:1 rank:1 indicates:1 bos:6 helpful:3 dependent:1 bienenstock:1 expand:1 overall:1 issue:1 ill:1 denoted:1 equal:1 having:2 identical:3 quantitatively:1 composed:2 attempt:1 detection:1 evaluation:1 mixture:6 sixty:1 extreme:2 nl:1 devoted:2 unless:1 haifa:1 minimal:2 increased:2 ollendorff:1 subset:12 predictor:1 technion:3 uniform:3 inadequate:1 combined:2 thanks:1 ie:16 physic:2 squared:2 central:1 resort:1 expert:1 leading:1 rescaling:1 yp:1 potential:1 nonoverlapping:1 summarized:2 explicitly:2 depends:3 later:1 view:1 try:1 option:1 dug:1 contribution:1 square:18 il:1 formed:3 variance:48 yield:2 ofthe:1 generalize:1 boiled:1 produced:1 unaffected:1 phys:1 ed:2 lengthy:1 obvious:1 couple:1 popular:2 massachusetts:1 recall:1 conversation:1 actually:1 back:1 ok:1 wesley:1 supervised:1 improved:2 rand:3 correlation:2 hand:1 ei:2 replacing:1 overlapping:7 quality:2 perhaps:1 believe:1 effect:10 normalized:1 unbiased:1 read:1 bkh:1 moore:6 noted:1 criterion:1 trying:1 complete:6 demonstrate:1 performs:1 recently:2 superior:5 common:2 ug:5 kinzel:3 jp:1 extend:1 approximates:1 qkd:1 refer:1 fk:2 similarly:2 longer:1 recent:1 scenario:2 route:1 arbitrarily:1 discussing:1 accomplished:1 seen:3 minimum:3 analyzes:1 eo:4 converge:1 monotonically:1 signal:2 ii:1 full:6 thermodynamic:3 multiple:3 calculation:5 specializing:1 regression:4 basic:2 expectation:3 normalization:2 separately:1 singular:3 source:1 posse:1 seem:1 split:2 identically:1 concerned:1 easy:2 independence:3 fit:1 reduce:2 idea:1 domany:1 y112:1 whether:2 expression:6 effort:1 wo:7 hessian:1 afford:1 passing:3 proceed:3 useful:2 detailed:1 clear:3 ylx:2 amount:1 category:1 ugr:1 reduced:1 meir:5 xij:1 estimated:1 disjoint:1 key:1 nevertheless:1 drawn:1 wasteful:1 btl:1 replica:1 undertaken:1 sum:2 year:1 inverse:2 separation:1 draw:1 scaling:1 pay:1 quadratic:1 aspect:1 attempting:3 xty:1 department:1 according:2 combination:12 beneficial:2 smaller:1 slightly:1 lid:3 making:1 modification:1 rev:1 heart:1 taken:1 equation:6 agree:1 remains:1 previously:1 turn:2 granger:4 hh:1 mind:1 addison:1 yare:1 observe:2 outweighs:1 neglect:1 unchanged:2 objective:1 question:2 quantity:3 strategy:1 dependence:1 usual:1 separate:1 berlin:1 manifold:1 trivial:1 reason:2 assuming:1 index:1 concomitant:2 balance:1 minimizing:4 scharf:3 unfortunately:1 difficult:2 potentially:1 statement:1 expense:2 rise:1 unknown:1 twenty:1 finite:8 t:1 situation:8 looking:1 precise:1 dc:1 arbitrary:1 community:1 bk:3 pair:1 namely:5 learned:1 able:1 beyond:1 usually:3 below:2 mismatch:1 regime:4 event:1 overlap:1 demanding:1 solvable:1 scheme:2 improve:1 woo:1 review:2 literature:2 acknowledgement:1 relative:1 asymptotic:1 fully:2 expect:1 generation:1 interesting:2 straighforward:1 degree:1 edf:2 principle:1 balancing:1 course:1 keeping:1 free:1 drastically:1 bias:36 allow:1 fall:1 emp:2 taking:1 distributed:2 van:1 curve:3 dimension:2 opper:4 computes:1 author:1 made:1 coincide:1 obtains:1 keep:1 global:1 conclude:2 xi:1 continuous:1 learn:2 obtaining:1 du:4 complex:2 necessarily:1 multilayered:1 noise:9 allowed:1 referred:1 hemmen:1 fashion:1 schulten:1 explicit:5 xl:1 late:1 third:1 down:1 theorem:1 rk:1 xt:3 specific:3 r2:1 decay:1 dominates:1 exists:1 albeit:1 effectively:3 magnitude:1 nk:2 forecast:2 simply:5 penrose:6 expressed:2 g2:6 u2:1 corresponds:1 goal:1 hard:1 springerverlag:1 typical:1 specifically:2 reducing:1 wt:1 called:2 perceptrons:1 support:1 latter:1 arises:2 meant:1 brevity:1 correlated:1 |
7,238 | 997 | A Real Time Clustering CMOS
Neural Engine
T. Serrano-Gotarredona, B. Linares-Barranco, and J. L. Huertas
Dept. of Analog Design, National Microelectronics Center (CNM), Ed. CICA, Av. Reina
Mercedes sIn, 41012 Sevilla, SPAIN. Phone: (34)-5-4239923, Fax: (34)-5-4624506,
E-mail: bernabe@cnm.us.es
Abstract
We describe an analog VLSI implementation of the ARTI algorithm
(Carpenter, 1987). A prototype chip has been fabricated in a standard low
cost 1.5~m double-metal single-poly CMOS process. It has a die area of
lcm2 and is mounted in a 12O-pins PGA package. The chip realizes a
modified version of the original ARTI architecture. Such modification
has been shown to preserve all computational properties of the original
algorithm (Serrano, 1994a), while being more appropriate for VLSI
realizations. The chip implements an ARTI network with 100 F1 nodes
and 18 F2 nodes. It can therefore cluster 100 binary pixels input patterns
into up to 18 different categories. Modular expansibility of the system is
possible by assembling an NxM array of chips without any extra
interfacing circuitry, resulting in an F1 layer with l00xN nodes, and an
F2 layer with 18xM nodes. Pattern classification is performed in less
than 1.8~s, which means an equivalent computing power of 2.2x109
connections and connection-updates per second. Although internally the
chip is analog in nature, it interfaces to the outside world through digital
signals, thus having a true asynchrounous digital behavior. Experimental
chip test results are available, which have been obtained through test
equipments for digital chips.
1
INTRODUCTION
The original ARTI algorithm (Carpenter, 1987) proposed in 1987 is a massively parallel
architecture for a self-organizing neural binary-pattern recognition machine. In response
to arbitrary orderings of arbitrarily many and complex binary input patterns, ARTI is
756
T. Serrano-Gotarredona, B. Linares-Barranco, 1. L. Huertas
Initialize weights:
zij
=1
N
L Z/i-LB L Zij+LM
Tj = LA
i
Read input pattern:
I = (/1' 12,
... I N )
N
=1
i =1
Winner-Take-All:
=
=
Y,
1 if T, maxj {Tj
y . = 0 if j:t;l
}
J
'-- - - - - - - - - - - - - - - - - - - l
z,1 new = In z,1 old
Fig.!: Modified Fast Learning or Type-3 ART!
implementation algorithm
capable of learning, in an unsupervised way, stable recognition codes. The ARTl
architecture is described by a set of Short Term Memory (STM) and another set of Long
Term Memory (LTM) time domain nonlinear differential equations. It is valid to assume
that the STM equations settle much faster (instantaneously) than the LTM equations, so
that the STM differential equations can be substituted by nonlinear algebraic equations
that describe the steady-state of the STM differential equations. Furthermore, in the
fast-learning mode (Carpenter, 1987), the LTM differential equations are as well
substituted by their corresponding steady-state nonlinear algebraic equations. This way,
the ARTI architecture can be behaviorally modelled by the sequential application of
nonlinear algebraic equations. Three different levels of ARTI implementations (both in
software and in hardware) can therefore be distinguished:
Type-I: Full Model Implementation: both STM and LTM time-domain differential
equations are realized. This implementation is the most expensive (both in software
and in hardware), and requires a large amount of computational power.
Type-2: STM steady-state Implementation: only the LTM time-domain differential
equations are implemented. The STM behavior is governed by nonlinear algebraic
equations. This implementation requires less resources than the previous one.
However, a proper sequencing of STM events has to be introduced artificially, which
is architecturally implicit in the Type-I implementation.
Type-3: Fast Learning Implementation: STM and LTM is implemented with algebraic
equations. This implementation is computationally the less expensive one. In this case
an artificial sequencing of STM and LTM events has to be done.
The implementation presented in this paper realizes a modified version of the original
ARTI Type-3 algorithm, more suitable for VLSI implementations. Such modified ARTI
system has been shown to preserve all computational properties of the original ARTI
architecture (Serrano, 1994a). The flow diagram that describes the modified ARTI
architecture is shown in Fig. I. Note that there is only one binary-valued weight template
(Zij)' instead ofthe two weight templates (one binary-valued and the other real-valued) of
the original ARTl. For a more detailed discussion of the modified ARTI algorithm refer to
(Serrano, 1994a, 1994b).
In the next Section we will provide an analog current-mode based circuit that implements
in hardware the flow diagram of Fig. 1. Note that, although internally this circuit is analog
in nature, from its input and output signals point of view it is a true asynchronous digital
757
A Real Time Clustering CMOS Neural Engine
circuit, easy to interface with any conventional digital machine. Finally, in Section 3 we
will provide experimental results measured from the chip using a digital data acquisition
test equipment.
2
CIRCUIT DESCRIPTION
The ART! chip reported in this paper has an F J layer with 100 neurons and an F2 layer
with 18 neurons. This means that it can handle binary input patterns of 100 pixels each,
and cluster them into up to 18 different categories; according to a digitally adjustable
vigilance parameter p. The circuit architecture of the chip is shown in Fig. 2(a). It consists
of an array of 18x100 synapses, a lx100 array of "vigilance synapses", a unity gain
18-outputs current mirror, an adjustable gain 18-outputs current mirror (with p=O.O, 0.1, ...
0.9)1, 18 current-comparator-controlled switches and an 18-input-currents
Winner-Take-All (WTA) (Serrano, 1994b). The inputs to the circuit are the 100 binary
digital input voltages Ii ' and the outputs of the circuit are the 18 digital output voltages Yj .
External control signals allow to change parameters p, LA' L B , and L M . Also, extra
circuitry has been added for reading the internal weights Zij while the system is learning.
Each row of synapses generates two currents,
100
100
Tj = LA LZ/i-LBLZij+LM
i= 1
i= 1
100
(1)
Vj = LA LZ/i
i
=1
while the row of the "vigilance synapses" generates the current
100
Vp = LALli
(2)
i= 1
Each of the current comparators compares the current V . versus pVp ' and allows current
T. to reach the WTA only if pV ~ V . . This way dompetition and vigilance occur
srmultaneously and in parallel, speeding tip significantly the search process.
Fig. 2(b) shows the content of a synapse in the 18x 100 array. It consists of three current
sources with switches, two digital AND gates and a flip-flop. Each synapse receives two
input voltages Ii and y . , and two global control voltages <1>/ (to enable/disable learning)
and reset (to initializ6 all weights zj" to '1'). Each synapse generates two currents
LAlizij-LBzij and LAliz ij ,which will be1summed up for all the synapses in the same row to
generate the currents T. and V.. If learning is enabled (<I> / = 1) the value of Zj" will
change to Iiz i . if y . = f . The ,lvigilance synapses" consist each of a current-sou~ce of
value L A witI~ a s~itch controlled by the input voltage Ii' The current comparators are
those proposed in (Dominguez-Castro, 1992), the WTA used is reported in (Lazzaro,
1989), and the digitally adjustable current mirror is based on (Loh, 1989), while its
continuous gain fine tuning mechanism has been taken from (Adams, 1991).
1. An additional pin ofthe chip can fine-tune
p between 0.9 and 1.0.
T. Serrano-Gotarredona, B. Linares-Barranco, J. L. Huertas
758
-------....... -------..... --- 1
1
,,
,,
:l,t.ti~j
11 12
.
.
.
1\00
(a)
(b)
Fig. 2: (a) System Diagram of Current-Mode ART! Chipt (b) Circuit Diagram of Synapse
Fig. 3: Tree based current-mirror scheme for matched current sources
The circuit has been designed in such a way that the WTA operates with a precision
around 1.5% (--6 bits). This means that all LA and L8 current sources have to match
within an error of less than that. From a circuit implementation point of view this is not
easy to achieve, since there are 5500 current sources spread over a die area of lcm2 .
Typical mismatch between reasonable size MOS transistors inside such an area extension
can be expected to be above 10% (pelgrom, 1989). To overcome this problem we
implemented a tree-based current mirror scheme as is shown in Fig. 3. Starting from a
unique current reference, and using high-precision lO(or less)-outputs current mirrors
(each yielding a precision around 0.2%), only up to four cascades are needed. This way,
the current mismatch attained at the synapse current sources was around 1% for currents
between LA/8 = 5J.LA and LA/8 = 10~A. This is shown in Fig. 4, where the measured dc
output current-error (in %) versus input current of the tree based configuration for 18 of
the 3600 LA synapse sources is depicted.
A Real Time Clustering CMOS Neural Engine
Fig. 4: Measured current mirror cascade missmatch (1 %/div) for LA for currents below
3
759
10~A
EXPERIMENTAL RESULTS
Fig. 5 shows a microphotograph of a prototype chip fabricated in a standard digital
double-metal, single-poly 1.5~m low cost CMOS process. The chip die area is Icm2 , and
it is mounted in a 120-pins PGA package. Fig. 6 shows a typical training sequence
accomplished by the chip and obtained experimentally using a test equipment for digital
chips. The only task performed by the test equipment was to provide the input data
patterns I (first column in Fig. 6), detect which of the output nodes became 'I' (pattern
with a vertical bar to its right), and extract the learned weights. Each IOxlO square in Fig.
6 represents either a lOa-pixels input vector I, or one row of lOa-pixels synaptic weights
z . == ezi" zp ... zIOO?) . Each row of squares in Fig. 6 represents the input pattern (first
square)land the 18 1vectors Zj after learning has been performed for this input pattern. The
sequence shown in Fig. 6 has been obtained for p = 0.7, LA = JO/lA, LB = 9.5/lA , and
LM = 950 /lA. Only two iterations of input patterns presentations were necessary, in this
case, for the system to learn and self-organize in response to these 18 input patterns.
The last row in Fig. 6 shows the final learned templates. Fig. 7 shows final learned
templates for different values of p. The numbers below each square indicate the input
patterns that have been clustered into each Zj category.
Delay time measurements have been performed for the feedforward action of the chip
(establishment of currents T., V. , and Vp ' and their competitions until the WTA settles),
and for the updating of weights: The feedforward delay is pattern and bias currents (LA'
L B , LM ) dependent, but has been measured to be always below 1.6/ls. The learning time
is constant and is around 180ns. Therefore, throughput time is less than 1.8/ls. A digital
neuroprocessor able to perform a connections/s, b connection-updates/s, and with a
dedicated WTA section with a c seconds delay, must satisfy
T. Serrano-Gotarredona. B. Linares-Barranco. J. L Huertas
760
Fig. 5: Microphotograph of ARTl chip
3700 + 100 + c = 1.8J..lS
a
b
(3)
to meet the performance of our ~rototype chip. If a = band c = lOOns, the equivalent
speed would be a = b = 2.2 x 10 connections and connection-updates per second.
4
CONCLUSIONS
A high speed analog current-mode categorizer chip has been built using a standard low
cost digital CMOS process. The high performance of the chip is achieved thanks to a
simplification of the original ARTl algorithm. The simplifications introduced are such that
all the original computational properties are preserved. Experimental chip test results are
provided.
A Real Time Clustering CMOS Neural Engine
Z, Z2 Z3 Z4 Zs z6 Z7 Zg Z9 Z10 Zll Z12 Z13 Z14 ZlS Z16 Z17 Zig
1?
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
13. ??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
?????? U???????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
??????????????????
2?
3.
4 ?
5 ?
6 ?
7 ?
8 ?
9?
10 ?
11 ?
12 ?
14 ?
15 ?
16 ?
17 ?
18 ?
1 ?
2 ?
3 ?
4 ?
5 ?
6 ?
7 ?
8 ?
9_
10 ?
11 ?
12 ?
13 ?
14
II
15 ?
16 ?
17 ?
18 ?
Fig. 6: Test sequence obtained experimentally for p=O.7, LA=lO/-lA, LB=9.5/-lA, and
LM=950/-lA
761
T. Serrano-Gotarredona, B. Linares-Barranco, J. L. Huertas
762
Fig. 7: Categorization of the input patterns for LA=3.2~A, LB=3.0~, LM=400~A, and
different values of p
References
W. J. Adams and J. Ramimez-Angulo. (1991) "Extended Transconductance AdjustmentlLinearisation
Technique," Electronics Letters, vol. 27, No. 10, pp. 842-844, May 1991.
G. A. Carpenter and S. Grossberg. (1987) "A Massively Parallel Architecture for a Self-Organizing
Neural Pattern Recognition Machine," Computer VIsion, Graphics, and Image Processing, vol. 37,
pp. 54-115, 1987.
R. Dominguez-castro, A. Rodrfguez-Vazquez, F. Medeiro, and 1. L. Huertas. (1992) "High
Resolution CMOS Current Comparators," Proc. of the 1992 European Solid-State Circuits
Conference (ESSCIRC'92), pp. 242-245, 1992.
J. Lazzaro, R. Ryckebusch, M. A. Mahowald, and C. Mead. (1989) "Winner-Take-All Networks of
O(n) Complexity," in Advances in Neural Information Processing Systems, vol. 1, D. S. Touretzky
(Ed.), Los Altos, CA: Morgan Kaufmann, 1989, pp. 703-711.
K. Loh, D. L. Hiser, W. J. Adams, and R. L. Geiger. (1989) "A Robust Digitally Programmable and
ReconfigurabJe Monolithic Filter Structure," Proc. of the 1989 Int. Symp. on Circuits and Systems
(ISCAS'89), Portland, Oregon, vol. 1, pp. 110-113, 1989.
M. J. Pelgrom, A. C. J. Duinmaijer, and A. P. G. Welbers. (1989) "Matching Properties of MOS
Transistors," IEEE Journal of Solid-State Circuits, vol. 24, No.5, pp. 1433-1440, October 1989.
T. Serrano-Gotarredona and B. Linares-Barranco. (1994a) "A Modified ARTl Algorithm more
suitable for VLSI Implementations," submitted for publication (journal paper).
T. Serrano-Gotarredona and B. Linares-Barranco. (1994b) "A Real-Time Clustering Microchip
Neural Engine," submitted for publication (journal paper).
| 997 |@word version:2 arti:12 solid:2 electronics:1 configuration:1 zij:4 current:35 z2:1 must:1 zll:1 designed:1 update:3 short:1 node:5 differential:6 consists:2 microchip:1 symp:1 inside:1 expected:1 behavior:2 z13:1 spain:1 stm:10 matched:1 provided:1 circuit:13 alto:1 z:1 fabricated:2 ti:1 control:2 internally:2 organize:1 monolithic:1 mead:1 meet:1 grossberg:1 unique:1 yj:1 implement:2 area:4 z7:1 significantly:1 cascade:2 matching:1 equivalent:2 conventional:1 center:1 starting:1 l:3 resolution:1 array:4 enabled:1 artl:5 handle:1 recognition:3 expensive:2 updating:1 ordering:1 zig:1 digitally:3 architecturally:1 cnm:2 complexity:1 f2:3 chip:21 x100:1 fast:3 describe:2 artificial:1 outside:1 modular:1 valued:3 final:2 sequence:3 transistor:2 reset:1 serrano:11 realization:1 organizing:2 achieve:1 description:1 competition:1 los:1 double:2 cluster:2 zp:1 categorization:1 cmos:8 adam:3 itch:1 measured:4 ij:1 implemented:3 indicate:1 filter:1 enable:1 settle:2 microphotograph:2 f1:2 clustered:1 extension:1 around:4 mo:2 lm:6 circuitry:2 proc:2 realizes:2 instantaneously:1 interfacing:1 behaviorally:1 always:1 modified:7 establishment:1 voltage:5 publication:2 sou:1 portland:1 sequencing:2 equipment:4 detect:1 dependent:1 vlsi:4 pixel:4 classification:1 art:3 initialize:1 having:1 represents:2 unsupervised:1 comparators:3 throughput:1 loon:1 preserve:2 national:1 maxj:1 iscas:1 yielding:1 tj:3 capable:1 necessary:1 tree:3 old:1 column:1 mahowald:1 cost:3 delay:3 graphic:1 reported:2 thanks:1 tip:1 jo:1 vigilance:4 external:1 int:1 z12:1 oregon:1 satisfy:1 z16:1 performed:4 view:2 parallel:3 square:4 became:1 kaufmann:1 ofthe:2 vp:2 reina:1 modelled:1 vazquez:1 submitted:2 synapsis:6 reach:1 touretzky:1 ed:2 synaptic:1 acquisition:1 pp:6 gain:3 z9:1 attained:1 response:2 synapse:6 done:1 furthermore:1 implicit:1 until:1 receives:1 nonlinear:5 mode:4 categorizer:1 true:2 read:1 linares:7 sin:1 self:3 die:3 steady:3 dedicated:1 interface:2 image:1 barranco:7 winner:3 analog:6 assembling:1 refer:1 measurement:1 tuning:1 z4:1 stable:1 ezi:1 phone:1 massively:2 binary:7 arbitrarily:1 accomplished:1 morgan:1 additional:1 disable:1 signal:3 ii:4 full:1 faster:1 match:1 long:1 controlled:2 vision:1 iteration:1 achieved:1 preserved:1 fine:2 diagram:4 source:6 extra:2 flow:2 feedforward:2 easy:2 switch:2 architecture:8 prototype:2 loh:2 algebraic:5 lazzaro:2 action:1 programmable:1 detailed:1 tune:1 amount:1 band:1 hardware:3 category:3 generate:1 zj:4 per:2 huertas:6 vol:5 four:1 mercedes:1 ce:1 ioxlo:1 package:2 letter:1 reasonable:1 geiger:1 x109:1 bit:1 pga:2 layer:4 simplification:2 esscirc:1 occur:1 software:2 generates:3 speed:2 transconductance:1 according:1 zls:1 describes:1 unity:1 wta:6 modification:1 dominguez:2 castro:2 sevilla:1 taken:1 computationally:1 equation:13 nxm:1 resource:1 pin:3 mechanism:1 needed:1 flip:1 available:1 z10:1 appropriate:1 distinguished:1 gate:1 original:8 clustering:5 added:1 realized:1 ryckebusch:1 div:1 mail:1 code:1 z3:1 october:1 design:1 implementation:14 proper:1 adjustable:3 perform:1 av:1 neuron:2 vertical:1 flop:1 extended:1 dc:1 arbitrary:1 lb:4 introduced:2 connection:6 engine:5 learned:3 able:1 bar:1 below:3 pattern:16 xm:1 mismatch:2 reading:1 built:1 memory:2 power:2 suitable:2 event:2 scheme:2 extract:1 fax:1 speeding:1 mounted:2 versus:2 digital:13 metal:2 land:1 row:6 l8:1 lo:2 loa:2 last:1 asynchronous:1 bias:1 allow:1 template:4 overcome:1 world:1 valid:1 lz:2 global:1 cica:1 gotarredona:7 search:1 continuous:1 z6:1 nature:2 iiz:1 learn:1 ca:1 robust:1 poly:2 complex:1 artificially:1 domain:3 vj:1 substituted:2 european:1 spread:1 carpenter:4 fig:21 n:1 precision:3 pv:1 governed:1 microelectronics:1 consist:1 sequential:1 ltm:7 mirror:7 depicted:1 comparator:1 presentation:1 content:1 change:2 experimentally:2 typical:2 operates:1 e:1 experimental:4 la:20 zg:1 internal:1 dept:1 |
7,239 | 998 | Learning direction in global motion: two
classes of psychophysically-motivated
models
V. Sundareswaran
Lucia M. Vaina*
Intelligent Systems Laboratory, College of Engineering,
Boston University
44 Cummington Street, Boston, MA 02215
Abstract
Perceptual learning is defined as fast improvement in performance and
retention of the learned ability over a period of time. In a set of psychophysical experiments we demonstrated that perceptual learning occurs for the discrimination of direction in stochastic motion stimuli. Here
we model this learning using two approaches: a clustering model that
learns to accommodate the motion noise, and an averaging model that
learns to ignore the noise. Simulations of the models show performance
similar to the psychophysical results.
1
Introduction
Global motion perception is critical to many visual tasks: to perceive self-motion,
to identify objects in motion, to determine the structure of the environment, and
to make judgements for safe navigation. In the presence of noise, as in random dot
kinematograms, efficient extraction of global motion involves considerable spatial
integration. Newsome and Colleagues (1989) showed that neurons in the macaque
middle temporal area (MT) are motion direction-selective, and perform global integration of motion in their large receptive fields. Psychophysical studies in humans
have characterized the limits of spatial and temporal integration in motion (Watamaniuk et. aI, 1984) and the nature of the underlying motion computations (Vaina
et. al 1990).
?Please address all correspondence to Lucia Vaina
918
V. Sundareswaran, Lucia M. Vaina
Since the psychophysical and neural substrate of global motion are fairly well understood, we were interested to see whether the perception of direction in such
global motion stimuli can improve with practice. Studies specifically addressing
this question for other early perceptual tasks have shown that improvements of performance obtained in the first experimental session are preserved in a subsequent
session and retained over weeks. This is considered as perceptual learning (Gibson,
1953). Psychophysical studies of perceptual learning show that the beneficial effects
of practice are lost if some stimulus parameters are changed significantly, such as
orientation, spatial frequency or location in the visual field. Based on the time scale
necessary for the improvement to occur, two major learning paradigms have been
used in perceptual learning : slow, progressive learning (several thousand trials are
required to reach stable performance) and fast !earning (improvement occurs and
stabilizes in the first 100-200 trials).
The idea of fast learning and the nature of its limits is attractive from a computational point of view because it encourages the exploration of practice-dependent
plasticity found in the adult early visual system (Fregnac et. al., 1988, Gilbert and
Wiesel 1992). A recent line of research in biologically motivated learning models
originated by Poggio (Poggio, 1990) takes perceptual learning as evidence that"the
brain may be able to synthesize-possibly in the cortex-appropriate task-specific
modules that receive input from retinotopic cells and learn to solve the task after
a short training phase in which they are exposed to examples of the task". Poggio
and colleagues (1992) have illustrated this approach in learning vernier hyperacuity.
Here, we adopted this general framework to study learning of direction in global
motion. In contrast to Poggio et. aI's supervised learning paradigm, we used unsupervised learning both in the psychophysical experiments and modeling of learning.
We designed a set of psychophysical tasks to study whether fast learning occurs in
discrimination of opposite directions of global motion and to explore the limits of
this learning. To model the learning, we studied two models that differ in the way
they deal with noise.
2
Psychophysics
Ball and Sekuler (1982, 1987) showed that discriminability of the direction of motion
of two random dot patterns improved with training. In this learning paradigm, more
than 2000 trials are required for reaching a stable performance. Such a "slow" learning time scale has been reported for the learning of other perceptual tasks, such as
vernier acuity (McKee and Westheimer 1978, Fahle 1994), stereoacuity (Fendick and
Westheimer, 1983; Ramachandran and Braddick 1973) and discrimination of line
orientation (Vogels and Orban 1985). In contrast to this "slow learning," Fiorentini
and Berardi (1981) showed that for learning the discrimination of complex gratings with two harmonics of different spatial phase, 100-200 trials suffice. Similarly
Poggio et. al. (1992) show that a small number of trials suffice for significantly
improving performance on a vernier hyperacuity task. Both studies discussed the
specificity of learning to the stimulus attributes.
In our study, we used a two-alternative, forced-choice psychophysical procedure to
measure the subject's ability to discriminate between two opposite directions of motion in dynamic random dot patterns in which 25% of the dots provide a correlated
Learning Direction in Global Motion
919
Figure 1: (a) Stimulus: a fraction of the dots (filled circles) move coherently in the
signal direction; the rest (open circles) move randomly, (b) Fast improvement is observed
on the first day of testing, and retained over a period of time. Each block consisted of
40 trials; data points averaged over 4 subjects, and errorbars show standard error.
motion signal spatially dispersed in a masking motion noise due to random motion
of the reminder 75% of dots (Fig. 1(a)). Each trial lasted 90 msec during which two
frames were presented (with inter-frame interval equal to zero). A session consisted
of 4-6 blocks of 40 trials each. Feedback was not provided during the experimental
sessions. Observers were required to maintain fixation on a fixation mark placed at
2? from the imaginary circumference of the stimulus.
To investigate the effects of practice and their retention, the discrimination of leftward vs rightward direction of motion in the display was tested first on each experimental session for 3 consecutive days and repeated 10 days later. The results are
presented in Fig. 1(b). For most observers, a fast and dramatic improvement was
seen in the first day. Fig. 1(b) shows that learning was maintained in subsequent
days, and even ten days later without any training in between. Examination of the
individual observers' data revealed that improvement of performance occurred only
if they started above chance level. This suggests that this fast learning might imply
the improvement of an existing representation of the stimulus.
In additional experiments, we did not find transfer to another direction of motion
(up/down), indicating that the learning is selective to specific characteristics of
the stimulus. Details of experiments testing the limits of the learning appear in
Vaina et. al (1995).
3
Modeling
We propose two paradigms to model the learning found in psychophysics. Both
use directionally-tuned units with properties similar to those of neurons found in
MT (Maunsell and Van Essen, 1983). Schematic MT neurons used in our modeling
integrate global information by summing over localized responses:
Xt
~ -(~(oi-od)
= ~e 2<1"
?
(1)
i=l
where (Jh is the standard deviation of the tuning, and information from n local
responses is taken into account. The responses of a collection of such units (each
920
V. Sundareswaran, Lucia M. Vaina
(a)
(b)
Figure 2: Architectures of the models: (a) Learning to Accommodate: Cluster Gaussians !I and h operate on input vector X to decide global motion direction, (b)
Learning to Ignore: Global motion direction is computed as a weighted combination of
units' preferred directions.
tuned to a different direction) form the input vector to the models.
3.1
Learning to accommodate
This model is based on grouping similar data. In this paradigm, after learning, the
network has an estimate of the contribution of noise, and takes it into account when
performing the task of discriminating between two different directions of motion;
so, we say that the model "learns to accommodate" the noise.
Fig 2(a) contains a schematic of this model. The representation vector consists of
responses Xl.X2, ..? , Xn of the directionally-tuned units .. Clustering is done in the
space of the representation vectors. The model is a combination of HyperBF-like
functions and clustering (Poggio et. aI, 1992, Moody and Darken 1989). We use
gaussians with mean at the cluster centers, a.nd "move" the cluster centers by a
learning algorithm.
A cluster gaussian computes a gaussian function of the representation vector for
the current stimulus from the current center of the corresponding cluster. We say
"current" because the center is moved as the learning proceeds. At any given point,
a center is at the current best estimate of the center of the corresponding data
cluster. More precisely, the learning rule to modify the jth coordinate of the center
is given by:
c(t-+:l) = c(t) . + '11 * (x~t) _ c(t) . )
(2)
W,)
W,)
'/
)
W,) ,
where w is the index of "winning" cluster and Cw,j is the jth coordinate of the
center for the wth cluster. This moves the center towards the new data vector X
that has been judged to belong to the wth cluster. The parameter TJ controls the
learning rate.
921
Learning Direction in Global Motion
Learning to ignore
3.2
The learning involved in this approach is Hebbian, and is termed "learning to ignore," because in this weighted averaging scheme, as learning occurs, the weights
for the noise response are progressively reduced to zero, leaving only the contribution from the signal. In other words, the network learns to ignore the noise (Vaina
et. al 1995).
The model's output is a global motion direction. If the weights associated with the
responses Xl, X2,? . ? Xn are WI, W2,'" W n , the global motion direction is calculated
as
II
Uo
=
tan
-1
(L:'"w
wiXisinOi )
WiXiCOSOj
'
where ti is the tuned direction (angle) of the ith unit. A schematic of the model is
shown in Fig. 2(b). The global direction is judged to be rightward if +Ot > 80 > -Ot.
We have examined two different learning rules: exposure-based learning, and selfsupervised learning.
Exposure-based learning: The weight corresponding to a unit is incremented by
an amount proportional to the current weight. Only units whose response values
are above a certain threshold are allowed to increase their weights. This learning
rule favors units that fire consistently:
(3)
where
rt
is a threshold, and
1]
is a small fraction that controls the learning rate.
Self-supervised learning: The weight corresponding to a unit is increased by an
amount proportional to the product of the current weight and a decreasing function (exponential) of the angular difference between the calculated global motion
direction and the direction of tuning of the unit:
Wi
~
Wi
+ 1]Wi e (- (O; -OO)2/ 2u e) ,
if
Xi> rt .
In this case, the model uses its own estimate of the global direction as an internal
feedback to determine the learning.
An approach similar to this model for learning vernier hyperacuity was proposed
by Weiss et. al (1993).
4
Experiments
For the simulations, the input (motion) vectors were represented by their angles relative to the positive horizontal axis (Le. , the magnitude is ignored). The responses of
the directionally-tuned units can be directly computed from the angles (see Eqn. 1;
alternatively, cosine tuning functions were used, and similar results were obtained) .
For each trial, the coherent motion direction was randomly decided. In the exper, where nd is the number of unit preferred directions; the
iments, O'h was set to ~
nd
directions Ot are chosen by uniformly dividing nd in to 211'. In all the experiments,
eight preferred directions were used; each trial contained 40 random dots moving
with a correlation of25% (the same correlation as in the tests with human subjects);
922
V. Sundareswaran, Lucia M. Vaina
1.00
0.9
I
u
0.90
0.8
0.80
- G - Espoeure
8
'f 0.70
0.7
0.60
0.6
~
Do
14
Block No
--.-- Self?npervised
O.S
0
s
10
1S
20
Block No
Figure 3: Results from typical simulation runs. On the left is the curve for the learning
to accommodate model (1} = 0.005; cluster (j = 0.5) and on the right are the curves
for the learning to ignore model.
each block consisted of 50 trials. The performance is measured by fraction correct,
corresponding to the fraction of the inputs that were correctly classified.
For both models, we did simulations to justify the architecture of the models by
disabling learning, and studying the performance of the models for increasing correlation. The performance was qUalitatively similar to that described in perceptual
studies of these stimuli in humans and monkeys.
Learning curves from simulations of both models are shown in Fig. 3. These are
results from averaging over the performance in ten simulation runs. As can be
seen, the models learn to do the discrimination. While both models successfully
learned to do the direction discrimination, there are quantitative differences in the
learning curves. The learning to accommodate paradigm improved very rapidly
(somewhat faster than the human subjects). The exposure-based learning rule for the
learning to ignore paradigm learned at a rate comparable with the human observers.
However, the final performance in this case was not very stable, and oscillated (not
shown), consistent with the observations of Weiss et. al (1993) in learning vernier
hyperacuity. The self-supervised rule, on the other hand, reached a stable level
of performance, but the learning was slower; the reason is that this learning rule
exhibits instability if a high value of learning rate (1}) is used.
If either model trained on inputs containing horizontal correlated motion was
later presented with inputs containing vertical correlated motion, the performance
dropped to pre-training levels. This non-transfer is consistent with psychophysical
results mentioned in Section 2.
5
Discussion
In this study, we focused on learning of global motion direction, and not on motion
perception. Motion perception is well-understood both physiologically and psy-
Learning Direction in Global Motion
923
chophysically. In a series of studies of neuronal correlates of the perceptual decision
of direction of motion in stochastic motion signals like the stimuli used here, together with studies of performance on these stimuli on monkeys with MT lesions,
Newsome and his collaborators (for a review, Newsome et. aI, 1989) provide strong
support for the hypothesis that perceptual judgements of motion direction are by
and large based on the directional signals carried by MT neurons.
Salzman and Newsome (1994) reported that in trained monkeys, the perception of
motion in stochastic noise is more likely mediated by a winner-take-all mechanism
than by a weighted averaging mechanism. Interestingly, our clustering model, after
learning, is behaviorally similar to the winner-take-all mechanism: the dominating
component of the representation results in the assignment to the closest cluster.
Our psychophysical studies clearly demonstrate that perceptual learning of global
motion direction occurs. While significant progress has been made to understand
the mechanisms and underlying circuitry of learning (Zohary et. al 1994) as yet
there is no satisfactory biologically explanation of how and where this learning may
occur (Zohary and Newsome 1994).
Our focus in this paper was on computational models of learning direction in global
motion. For this, we proposed two approaches, which differ in the way they deal
with noise: one learns to accommodates noise, and the other learns to ignore it. We
do not advocate that one or the other of the models we proposed here provides the
biologically correct choice for this task. However, together with the psychophysics
described here these models suggest new experiments which we are now exploring
both psychophysically and computationally.
We hope that by closely connecting models and psychophysics while keeping in
mind the aim of neuronal compatibility, we will make progress in understanding
how the cortex learns so fast to discriminate direction of motion in extremely noisy
situations.
Acknowledgments
This research was conducted at the Intelligent Systems Laboratory of Boston University, College of Engineering. LMV and VS were supported in part by grants
from the Office of Naval Research (# N00014-93-1-0381) and the National Institute of Health (EY R01- 07861) to LMV. VS was in part supported by the Boston
University College of Engineering Dean's postdoctoral fellowship. We gratefully
acknowledge additional financial support for VS from the Dean's special research
fund.
References
[1] K. Ball and R. Sekuler. Direction-specific improvement in motion discrimination.
Vision Research, 27(6):953-965, 1987.
[2] M. Fahle. Human pattern recognition: parallel processing and perceptual learning.
Perception, 23:411-427, 1994.
[3] M. Fendick and G. Westheimer. Effects of practice and the separation of test targets
on foveal and perifoveal hyperacuity. Vision Research, 23: 145-150, 1983.
924
V. Sundareswaran, Lucia M. Vaina
[4] A. Fiorentini and N. Berardi. Perceptual learning specific for orientation and spatial
frequency. Nature, 287(5777):43-44, September 1980.
[5] Y. Fregnac, D. Shulz, S. Thorpe, and E. Bienstock. A cellular analogue of visual
cortical plasticity. Nature, 333:367-370, 1988.
[6] E. J. Gibson. Improvements in perceptual judgement as a function of controlled
practice of training. Psychology bulletin, 50:402-431, 1953.
[7] C. D. Gilbert and T. N. Wiesel. Receptive field dynamics in adult primary visual
cortex. Nature, 356:150-152, 1992.
[8] A. Karni and D. Sagi. Where practice makes perfect in texture discrimination: Evidence for primary visual cortex plasticity. Proc. Natl. Acad. Sci. USA, 88:4966- 4970,
June 1991.
[9] J. H. R. Maunsell and D. C. Van Essen. Functional properties of neurons in the middle
temporal visual area of the macaque monkey i: selectivity for stimulus direction,
speed, and orientation. J. Neurophysiology, 49:1127-1147, 1983.
[10] S. P. McKee and G. Westheimer. Improvement in venier acuity with practice. Perception ?3 Psychophysics, 24:258-262, 1978.
[11] W. T. Newsome, K. H. Britten, and J. A. Movshon. Neuronal correlates of a perceptual decision. Nature, 341 :52-54, 1989.
[12] T. Poggio. A theory of how the brain might work. In Cold Spring Harbor Symposia
on Quantitative Biology, pages 899-910. Cold Spring Harbor Laboratory Press, 1990.
[13] T. Poggio, M. Fahle, and S. Edelman. Fast perceptual learning in visual hyperacuity.
Science, 256:1018-1021, 1992.
[14] T. Poggio and F. Girosi. A theory of networks for approximation and learning. AI
Memo 1140, M.LT, July 1989.
[15] V. S. Ramachandran and O. Braddick. Orientation-specific learning in stereopsis.
Perception, 2:371-376, 1973.
[16] C. D. Salzman and W. T. Newsome. Neural mechanisms for forming a perceptual
decision. Science, 264:231-237, 1994.
[17] L. M. Vaina, V. Sundareswaran, and J. Harris. Computational learning and natural
learning. Cognitive Brain Research, 1995. (in press).
[18] L. M. Vaina, N. M. Grzywacz, and M. LeMay. Structure from motion with impaired
local-speed and global motion-field computations. Neuml Computation, 2:420-435,
1990.
[19] R. Vogels and G . A. Orban. The effect of practice on the oblique effect in line
orientation judgements. Vision Research, 25:1679-1687, 1985.
[20] S. Watamaniuk, R. Sekuler, and D. Williams. Direction perception in complex dynamic displays: the integration of direction information. Vision Research, 24:55- 62,
1984.
[21] Y. Weiss, S. Edelman, and M. Fahle. Models of perceptual learning in vernier hyperacuity. Neural Computation, 5:695-718, 1993.
[22] E. Zohary, S. Celebrini, K. H. Britten, and W. T . Newsome. Neuronal plasticity that
underlies improvement in perceptual performance. Science, 263:1289-1292, March
1994.
[23] E. Zohary and W. T. Newsome. Perceptual learning in a direction discrimination task
is not based upon enhanced neuronal sensitivity in the sts. Investigative Ophthalmology and Visual Science supplement, page 1663, 1994.
| 998 |@word neurophysiology:1 trial:11 middle:2 judgement:4 wiesel:2 nd:4 open:1 simulation:6 dramatic:1 accommodate:6 contains:1 series:1 foveal:1 tuned:5 interestingly:1 imaginary:1 existing:1 current:6 od:1 yet:1 subsequent:2 plasticity:4 girosi:1 designed:1 selfsupervised:1 progressively:1 fund:1 discrimination:10 v:4 stereoacuity:1 ith:1 short:1 oblique:1 wth:2 provides:1 location:1 symposium:1 edelman:2 consists:1 fixation:2 advocate:1 inter:1 brain:3 decreasing:1 zohary:4 increasing:1 provided:1 retinotopic:1 underlying:2 suffice:2 monkey:4 temporal:3 quantitative:2 ti:1 control:2 unit:12 maunsell:2 uo:1 appear:1 grant:1 positive:1 retention:2 dropped:1 engineering:3 sagi:1 understood:2 local:2 limit:4 acad:1 modify:1 fiorentini:2 might:2 discriminability:1 studied:1 examined:1 suggests:1 sekuler:3 averaged:1 decided:1 acknowledgment:1 sundareswaran:6 testing:2 practice:9 lost:1 block:5 procedure:1 cold:2 area:2 gibson:2 significantly:2 word:1 pre:1 specificity:1 suggest:1 judged:2 instability:1 gilbert:2 dean:2 demonstrated:1 circumference:1 center:9 exposure:3 williams:1 focused:1 oscillated:1 vaina:11 perceive:1 rule:6 his:1 financial:1 lemay:1 coordinate:2 grzywacz:1 target:1 tan:1 enhanced:1 substrate:1 us:1 hypothesis:1 synthesize:1 hyperacuity:7 recognition:1 observed:1 module:1 thousand:1 incremented:1 mentioned:1 environment:1 dynamic:3 trained:2 exposed:1 upon:1 rightward:2 represented:1 forced:1 fast:9 investigative:1 whose:1 solve:1 dominating:1 say:2 ability:2 favor:1 noisy:1 final:1 directionally:3 propose:1 product:1 rapidly:1 moved:1 cluster:11 impaired:1 perfect:1 object:1 oo:1 measured:1 progress:2 disabling:1 strong:1 dividing:1 grating:1 involves:1 berardi:2 differ:2 direction:42 safe:1 closely:1 correct:2 attribute:1 stochastic:3 exploration:1 human:6 venier:1 collaborator:1 exploring:1 considered:1 week:1 circuitry:1 stabilizes:1 major:1 early:2 consecutive:1 proc:1 successfully:1 weighted:3 hope:1 clearly:1 behaviorally:1 gaussian:2 aim:1 reaching:1 office:1 acuity:2 focus:1 naval:1 improvement:12 consistently:1 june:1 lasted:1 contrast:2 psy:1 dependent:1 selective:2 interested:1 compatibility:1 orientation:6 spatial:5 integration:4 fairly:1 psychophysics:5 special:1 field:4 equal:1 extraction:1 biology:1 progressive:1 unsupervised:1 stimulus:13 intelligent:2 thorpe:1 shulz:1 randomly:2 national:1 individual:1 phase:2 fire:1 maintain:1 investigate:1 essen:2 navigation:1 tj:1 natl:1 necessary:1 poggio:9 filled:1 circle:2 increased:1 modeling:3 newsome:9 assignment:1 addressing:1 deviation:1 conducted:1 reported:2 psychophysically:2 sensitivity:1 discriminating:1 together:2 connecting:1 moody:1 fregnac:2 containing:2 possibly:1 cognitive:1 account:2 sts:1 later:3 view:1 observer:4 reached:1 parallel:1 masking:1 contribution:2 oi:1 characteristic:1 identify:1 directional:1 classified:1 reach:1 colleague:2 frequency:2 involved:1 associated:1 reminder:1 supervised:3 day:6 response:8 improved:2 wei:3 done:1 angular:1 correlation:3 ramachandran:2 eqn:1 horizontal:2 hand:1 vogels:2 usa:1 effect:5 consisted:3 spatially:1 laboratory:3 satisfactory:1 illustrated:1 deal:2 attractive:1 during:2 self:4 encourages:1 please:1 maintained:1 cosine:1 cummington:1 demonstrate:1 motion:48 harmonic:1 mckee:2 mt:5 functional:1 celebrini:1 winner:2 discussed:1 occurred:1 belong:1 significant:1 ai:5 tuning:3 session:5 similarly:1 gratefully:1 dot:7 moving:1 stable:4 cortex:4 closest:1 own:1 showed:3 recent:1 leftward:1 termed:1 selectivity:1 certain:1 n00014:1 seen:2 additional:2 somewhat:1 ey:1 determine:2 paradigm:7 period:2 signal:5 ii:1 july:1 hebbian:1 faster:1 characterized:1 controlled:1 schematic:3 underlies:1 vision:4 cell:1 preserved:1 receive:1 fellowship:1 interval:1 leaving:1 w2:1 rest:1 operate:1 ot:3 subject:4 presence:1 revealed:1 harbor:2 psychology:1 architecture:2 opposite:2 idea:1 whether:2 motivated:2 movshon:1 ignored:1 amount:2 ten:2 reduced:1 correctly:1 threshold:2 fraction:4 run:2 angle:3 decide:1 separation:1 earning:1 decision:3 comparable:1 display:2 correspondence:1 occur:2 precisely:1 x2:2 lucia:6 speed:2 orban:2 extremely:1 spring:2 performing:1 watamaniuk:2 ball:2 combination:2 march:1 beneficial:1 ophthalmology:1 wi:4 biologically:3 taken:1 computationally:1 mechanism:5 mind:1 adopted:1 studying:1 gaussians:2 eight:1 appropriate:1 alternative:1 slower:1 neuml:1 clustering:4 r01:1 psychophysical:10 move:4 question:1 coherently:1 occurs:5 receptive:2 primary:2 rt:2 exhibit:1 september:1 cw:1 sci:1 braddick:2 street:1 accommodates:1 cellular:1 reason:1 salzman:2 lmv:2 retained:2 index:1 westheimer:4 vernier:6 memo:1 perform:1 vertical:1 neuron:5 darken:1 observation:1 acknowledge:1 situation:1 frame:2 hyperbf:1 required:3 errorbars:1 learned:3 coherent:1 macaque:2 address:1 adult:2 able:1 proceeds:1 perception:9 pattern:3 explanation:1 analogue:1 critical:1 natural:1 examination:1 scheme:1 improve:1 imply:1 axis:1 started:1 carried:1 mediated:1 britten:2 health:1 kinematograms:1 review:1 understanding:1 relative:1 proportional:2 localized:1 integrate:1 consistent:2 changed:1 placed:1 supported:2 keeping:1 jth:2 jh:1 understand:1 institute:1 bulletin:1 van:2 karni:1 feedback:2 calculated:2 xn:2 curve:4 cortical:1 computes:1 collection:1 qualitatively:1 made:1 correlate:2 ignore:8 preferred:3 fahle:4 global:23 summing:1 xi:1 alternatively:1 postdoctoral:1 stereopsis:1 physiologically:1 nature:6 learn:2 transfer:2 exper:1 improving:1 complex:2 did:2 noise:12 repeated:1 allowed:1 lesion:1 neuronal:5 fig:6 slow:3 originated:1 msec:1 winning:1 xl:2 exponential:1 iments:1 perceptual:21 learns:7 down:1 specific:5 xt:1 evidence:2 grouping:1 supplement:1 texture:1 magnitude:1 boston:4 lt:1 explore:1 likely:1 forming:1 visual:9 contained:1 chance:1 dispersed:1 harris:1 ma:1 towards:1 considerable:1 specifically:1 typical:1 uniformly:1 averaging:4 justify:1 discriminate:2 experimental:3 indicating:1 college:3 internal:1 mark:1 support:2 tested:1 correlated:3 |
7,240 | 999 | Correlation and Interpolation Networks for
Real-time Expression Analysis/Synthesis.
Trevor Darrell, Irfan Essa, Alex Pentland
Perceptual Computing Group
MIT Media Lab
Abstract
We describe a framework for real-time tracking of facial expressions
that uses neurally-inspired correlation and interpolation methods. A
distributed view-based representation is used to characterize facial state,
and is computed using a replicated correlation network. The ensemble
response of the set of view correlation scores is input to a network based
interpolation method, which maps perceptual state to motor control states
for a simulated 3-D face model. Activation levels of the motor state
correspond to muscle activations in an anatomically derived model. By
integrating fast and robust 2-D processing with 3-D models, we obtain a
system that is able to quickly track and interpret complex facial motions
in real-time.
1 INTRODUCTION
An important task for natural and artificial vision systems is the analysis and interpretation
of faces. To be useful in interactive systems and in other settings where the information
conveyed is of a time critical nature, analysis of facial expressions must occur quickly, or be
oflittle value. However, many of the traditional computer vision methods for estimating and
modeling facial state have proved difficult to perform fast enough for interactive settings.
We have therefore investigated neurally inspired mechanisms for the analysis of facial
expressions. We use neurally plausible distributed pattern recognition mechanisms to make
fast and robust assessments of facial state, and multi-dimensional interpolation networks to
connect these measurements to a facial model.
There are many potential applications of a system for facial expression analysis. Person-
910
Trevor Darrell. Irfan Essa. Alex Pentland
alized interfaces which sense a users emotional state, ultra-low bitrate video conferencing
which sends only facial muscle activations, as well as the enhanced recognition systems
mentioned above. We have focused on a application in computer graphics which stresses
both the analysis and synthesis components of our system: interactive facial animation.
In the next sections we develop a computational framework for neurally plausible expression
analysis, and the connection to a physically-based face model using a radial basis function
method. Finally we will show the results of these methods applied to the interactive
animation task, in which an computer graphics model of a face is rendered in real time, and
matches the state of the users face as sensed through a conventional video camera.
2
EXPRESSION MODELINGffRACKING
The modeling and tracking of expressions and faces has been a topic of increasing interest
recently. In the neural network field, several successful models of character expression
modeling have been developed by Poggio and colleagues. These models apply multidimensional interpolation techniques, using the radial basis function method, to the task
of interpolating 2D images of different facial expression. Librande [4] and Poggio and
Brunelli [9] applied the Radial Basis Function (RBF) method to facial expression modeling, using a line drawing representation of cartoon faces. In this model a small set of
canonical expressions is defined, and intermediate expressions constructed via the interpolation technique. The representation used is a generic "feature vector", which in the case of
cartoon faces consists of the contour endpoints. Recently, Beymer et al. [1] extended this
approach to use real images, relying on optical flow and image warping techniques to solve
the correspondence and prediction problems, respectively.
RBF-based techniques have the advantage of allowing for the efficient and fast computation of intermediate states in a representation. Since the representation is simple and
the interpolation computation straight-forward, real-time implementations are practical on
conventional systems. These methods interpolate between a set of 2D views, so the need for
an explicit 3-D representation is sidestepped. For many applications, this is not a problem,
and may even be desirable since it allows the extrapolation to "impossible" figures or expressions, which may be of creative value. However, for realistic rendering and recognition
tasks, the use of a 3-D model may be desirable since it can detect such impossible states.
In the field of computer graphics, much work has been done on on the 3-D modeling of
faces and facial expression. These models focus on the geometric and physical qualities
of facial structure. Platt and B adler [7], Pieper [6], Waters [11] and others have developed
models of facial structure, skin dynamics, and muscle connections, respectively, based on
available anatomical data. These models provide strong constraints for the tracking of
feature locations on a face. Williams et. al. [12] developed a method in which explicit
feature marks are tracked on a 3-D face by use of two cameras. Terzopoulos and Waters [10]
developed a similar method to track linear facial features, estimate corresponding parameters
of a three dimensional wireframe face model, and reproduce facial expression. A significant
limitation of these systems is that successful tracking requires facial markings. Essa and
Pentland [3] applied optical flow methods (see also Mase [5]) for the passive tracking of
facial motion, and integrated the flow measurement method into a dynamic system model.
Their method allowed for completely passive estimation of facial expressions, using all the
constraints provided by a full 3-D model of facial expression.
Both the view based method of Beymer et. al. and the 3-D model of Essa and Pentland rely
Correlation and Interpolation Networks for Real-Time Expression Analysis/Synthesis
(
911
(b)
Figure 1: (a) Frame of video being processed to extract view model. Outlined rectangle
indicates area of image used for model. (b) View models found via clustering method on
training sequence consisting of neutral, smile, and surprise expressions.
on estimates of optic flow, which are difficult to compute reliably, especially in real-time.
Our approach here is to combine interpolated view-based measurements with physically
based models, to take advantage of the fast interpolation capability of the RBF and the
powerful constraints imposed by physically based models. We construct a framework in
which perceptual states are estimated from real video sequences and are interpolated to
control the motor control states of a physically based face model.
3
VIEW-BASED FACE PERCEPTION
To make reliable real-time measurements of a complex dynamic object, we use a distributed
representation corresponding to distinct views of that object. Previously, we demonstrated
the use of this type of representation for the tracking and recognition of hand gestures [2].
Like faces, hands are complex objects with both non-rigid and rigid dynamics. Direct use
of a 3-D model for recognition has proved difficult for such objects, so we developed a
view-based method for representation. Here we apply this technique to the problem of facial
representation, but extend the scheme to connect to a 3-D model for high-level modeling
and generation/animation. With this, we gain the representational power and constraints
implied by the 3-D model as a high-level representation; however the 3-D model is only
indirectly involved in the perception stage, so we can still have the same speed and reliability
afforded by the view-based representation.
In our method each view characterizes a particular aspect or pose of the object being
represented. The view is stored iconically, that is, it is a literal image or template (but
with some point-wise statistics) of the appearance of the object in that aspect or pose. A
match criteria is defined between views and input images; usually a normalized correlation
function is used, but other criteria are possible. An input image is represented by the
ensemble of match scores from that image to the stored views.
To achieve invariance across a range of transformations, for example translation, rotation
and/or scale, units which compute the match score for each view are replicated at different
values of each transformation. I The unit which has maximal response across all values of
the transformation is selected, and the ensemble response of the view units which share the
1 In a computer implementation this exhaustive sampling may be impractical due to the number of units needed, in which case this stage may be approximated by methods which are hybrid
sampling/search methods.
912
Trevor Darrell, Irfan Essa, Alex Pentland
same transformation values as the selected unit is stored as the representation for the input
image. We set the perceptual state X to be a vector containing this ensemble response.
If the object to be represented is fully known a priori, then methods to generate views can
be constructed by analysis of the aspect graph if the object is polyhedral, or in general
by rendering images of the object at evenly spaced rotations. However, in practice good
3-D models that are useful for describing image intensity values are rare2 , so we look to
data-driven methods of acquiring object views.
As described in [2] a simple clustering algorithm can find a set of views that "span" a
training sequence of images, in the sense that for each image in the sequence at least one
view is within some threshold similarity to that image. The algorithm is as follows. Let
V be the current set of views for an object (initially one view is specified manually). For
each frame I of a training sequence, if at least one v E V has a match value M (v, 1) that is
greater than a threshold (J, then no action is performed and the next frame is processed. If
no view is close, then I is used to construct a new view which is added to the view set. A
view v' is created using a window of I centered at the location in the previous image where
the closest view was located. (All views usually share the same window size, determined
by the initial view.) The view set is then augmented to include the new view: V = V u v'.
This algorithm will find a set of views which well-characterizes an object across the range
of poses or expressions contained in the training sequence. For example, in the domain of
hand gestures, inputing a training sequence consisting of a waving hand will yield views
which contain images of the hand at several different rotations. In the domain of faces,
when input a training sequence consisting of a user performing 3 different expressions,
neutral, smile, and surprise, this algorithm (with normalized correlation and (J = 0.7) found
three views corresponding to these expressions to represent the face, as shown in Figure
l(b). These 3 views serve as a good representation for the face of this user as long as his
expression is similar to one in the training set.
The major advantage of this type of distributed view-based representation lies in the reduction of the dimensionality of the processing that needs to occur for recognition, tracking, or
control tasks. In the gesture recognition domain, this dimensionality reduction allowed for
conventional recognition strategies to be applied successfully and in real-time, on examples
where it would have been infeasible to evaluate the recognition criteria on the full signal. In
the domain explored in this paper it makes the interpolation problem of much lower order:
rather than interpolate from thousands of input dimensions as would be required when the
input is the image domain, the view domain for expression modeling tasks typically has on
the order of a dozen dimensions.
4
3-D MODELINGIMOTOR CONTROL
To model the structure of the face and the dynamics of expression performance, we use the
physically based model of Essa et. al. This model captures how expressions are generated
by muscle actuations and the resulting skin and tissue deformations. The model is capable
of controlled nonrigid deformations of various facial regions, in a fashion similar to how
humans generate facial expressions by muscle actuations attached to facial tissue. Finite
Element methods are used to model the dynamics of the system.
2 As opposed to modeling forces and shape deformations, for which 3-D models are useful and
indeed are used in the method presented here.
Correlation and Interpolation Networks for Real-Time Expression Analysis/Synthesis
9 J3
(a)
900.
1
8 0<5'> core
(b)
700
l\I1odel
ParalTleters
.
o0.2ActuatIons
(e)
-0.2
(d)
Figure 2: (a) Face images used as input, (b) normalized correlation scores X(t) for each
view model, (c) resulting muscle control parameters Y(t), (d) rendered images of facial
model corresponding to muscle parameters.
914
Trevor Darrell, Irfan Essa, Alex Pentland
This model is based on the mesh developed by Platt and Badler [7], extended into a
topologically invariant physics-based model through the addition of a dynamic skin and
muscle model [6, 11]. These methods give the facial model an anatomically-based facial
structure by modeling facial tissue/skin, and muscle actuators, with a geometric model to
describe force-based deformations and control parameters.
The muscle model provides us with a set of control knobs to drive the facial state, defined
to be a vector Y. These serve to define the motor state of the animated face. Our task now
is to connect the perceptual states of the observed face to these motor states.
5
CONNECTING PERCEPTION WITH ACTION
We need to establish a mapping from the perceptual view scores to the appropriate muscle
activations on the 3-D face model. To do this, we use multidimensional interpolation
strategies implemented in network form.
Interpolation requires a set of control points or exemplars from which to derive the desired
mapping. Example pairs of real faces and model faces for different expressions are presented
to the interpolation method during a training phase. This can be done in one of two ways,
with either a user-driven or model-driven paradigm. In the model-driven case the muscle
states are set to generate a particular expression by an animator/programmer and then the
user is asked to make the equivalent expression. The resulting perceptual (view-model)
scores are then recorded and paired with the muscle activation levels. In the user-driven
case, the user makes an expression of hislher own choosing, and the optic flow method
of Essa et. al. is used to derive the corresponding muscle activation levels. The modeldriven paradigm is simpler and faster, but the user-driven paradigm yields more detailed
and authentic facial expressions.
We use the Radial Basis Function (RBF) method presented in [8], and define the interpolated
motor controls to be a weighted sum of radial functions centered at each example:
n
Y = I:ci9(X - Xi)
(1)
i=1
where Y are the muscle states, X are the observed view-model scores, Xi are the example
scores, 9 is an RBF (and in our case was simply a linear ramp 9(?) = II?II), and the weights
Ci are computed from the example motor values Yi using the pseudo-inverse method [8].
6 INTERACTIVE ANIMATION SYSTEM
The correlation network, RBF interpolator, and facial model described above have been
combined into a single system for interactive animation. The entire system can be updated
at over 5 Hz, using a dedicated single board accelerator to compute the correlation network,
and an SGI workstation to render the facial mesh. ? Here we present two examples of the
processing performed by the system, using different strategies for coupling perceptual and
motor state.
Figure 2 illustrates one example of real-time facial expression tracking using this system,
using a full-coupling paradigm. Across the top, labeled (a), are five frames of a video
sequence of a user making a smile expression. This was one of the expressions used in the
training sequence for the view models shown in Figure 1(b), so they were applicable to be
Correlation and Interpolation Networks for Real-Time Expression Analysis/Synthesis
(a)
915
(b)
Figure 3: (a) Processing of video frame with independent view model regions for eyes,
eye-brows, and mouth region. (b) Overview shot of full system. User is on left, vision
system and camera is on right, and animated face is in the center of the scene. The animated
face matches the state of the users face in real-time, including eye-blinks (as is the case in
this shot.)
used here. Figure 2(b) shows the correlation scores computed for each of the 3 view models
for each frame of the sequence. This constituted the perceptual state representation, X(t).
In this example the full face is coupled with the full suite of motor control parameters. An
RBF interpolator was trained using perceptual/motor state pairs for three example full-face
expressions (neutral, smile, surprise); the resulting (interpolated) motor control values,
yet), for the entire sequence are shown in Figure 2(c). Finally, the rendered facial mesh for
five frames of these motor control values is shown in Figure 2(d).
When there are only a few canonical expressions that need be tracked/matched, this full-face
template approach is robust and simple. However if the user wishes to exercise independent
control of the various regions of the face, then the full coupling paradigm will be overly
restrictive. For example, if the user trains two expressions, eyes closed and eyes open, and
then runs the system and attempts to blink only one eye, the rendered face will be unable to
match it. (In fact closing one eye leads to the rendered face half-closing both eyes.)
A solution to this is to decouple the regions of the face which are independent geometrically
(and to some degree, in terms of muscle effect.) Under this paradigm, separate correlation networks are computed for each facial regions, and multiple RBF interpolations are
performed for each system. Each interpolator drives a distinct subset of the motor state
vector. Figure 3(a) shows the regions used for decoupled local templates. In these examples
independent regions were used for each eye, eyebrow, and the mouth region.
Finally, figure 3 (b) shows a picture of the set-up of the system as it is being run in an
interactive setting. The animated face mimics the facial state of the user, matching in real
time the position of the eyes, eyelids, eyebrows and mouth of the user. In the example
shown in this picture, the users eyes are closed, so the animated face's eyes are similarly
closed. Realistic performance of animated facial expressions and gestures are are possible
916
Trevor Darrell, Irfan Essa, Alex Pentland
through this method, since the timing and levels of the muscle activations react immediately
to changes in the users face.
7
CONCLUSION
We have explored the use of correlation networks and Radial Basis Function techniques for
the tracking of real faces in video sequences. A distributed view-based representation is
computed using a network of replicated normalized correlation units, and offers a fast and
robust assesment of perceptual state. 3-D constraints on facial shape are achieved through
the use of a an anatomically derived facial model, whose muscle activations are controled
via interolated perceptual states using the RBF method.
With this framework we have been able to acheive the fast and robust analysis and synthesis
of facial expressions. A modeled face mimics the expression of a user in real-time, using
only a conventional video camera sensor and no special marking on the face of the user. This
system has promise as a new approach in the interactive animation, video tele-conferencing,
and personalized interface domains.
References
[1] D. Beymer, A. Shashua, and T. Poggio, Example Based Image Analysis and Synthesis,
MIT AI Lab TR-1431, 1993.
[2] T. Darrell and A. Pentland. Classification of Hand Gestures using a View-Based
Distributed Representation In NIPS-6, 1993.
[3] I. Essa and A. Pentland. A vision system for observing and extracting facial action
parameters. In Proc. IEEE Conf. Computer Vision and Pattern Recognition, 1994.
[4] S. Librande. Example-based Character Drawing. S.M. Thesis, Media Arts and SciencelMedia Lab, MIT. 1992
[5] K. Mase. Recognition of facial expressions for optical flow. IEICE Transactions,
Special Issue on Computer Vision and its Applications, E 74( 10), 1991.
[6] S. Pieper, J. Rosen, and D. Zeltzer. Interactive graphics for plastic surgery: A task
level analysis and implementation. Proc. Siggraph-92, pages 127-134, 1992.
[7] S. M. Platt and N. I. Badler. Animating facial expression. ACM SIGGRAPH Conference Proceedings, 15(3):245-252,1981.
[8] T. Poggio and F. Girosi. A theory of networks for approximation and learning. MIT
AI Lab TR-1140, 1989.
[9] T. Poggio and R. Brunelli, A Novel Approach to Graphics, MIT AI Lab TR- 1354.
1992.
[10] D. Terzopoulus and K. Waters. Analysis and synthesis offacial image sequences using
physical and anatomical models. IEEE Trans. PAMI, 15(6):569-579, June 1993.
[11] K. Waters and D. Terzopoulos. Modeling and animating faces using scanned data.
The Journal of Visualization and Computer Animation, 2: 123-128, 1991.
[12] L. Williams. Performance-driven facial animation. ACM SIGGRAPH Conference
Proceedings, 24(4):235-242, 1990.
| 999 |@word open:1 sensed:1 tr:3 shot:2 reduction:2 initial:1 score:9 animated:6 current:1 activation:8 yet:1 must:1 mesh:3 realistic:2 girosi:1 shape:2 motor:13 half:1 selected:2 core:1 provides:1 location:2 simpler:1 five:2 constructed:2 direct:1 consists:1 combine:1 polyhedral:1 indeed:1 multi:1 inspired:2 relying:1 animator:1 window:2 increasing:1 provided:1 estimating:1 matched:1 medium:2 developed:6 transformation:4 impractical:1 suite:1 pseudo:1 brow:1 multidimensional:2 interactive:9 platt:3 control:14 unit:6 local:1 timing:1 interpolation:16 pami:1 range:2 practical:1 camera:4 practice:1 area:1 matching:1 integrating:1 radial:6 bitrate:1 close:1 impossible:2 conventional:4 map:1 imposed:1 demonstrated:1 equivalent:1 center:1 williams:2 focused:1 immediately:1 react:1 his:1 updated:1 enhanced:1 user:20 us:1 element:1 recognition:11 approximated:1 located:1 labeled:1 observed:2 capture:1 thousand:1 region:9 mentioned:1 asked:1 dynamic:7 trained:1 serve:2 basis:5 completely:1 siggraph:3 represented:3 various:2 train:1 distinct:2 fast:7 describe:2 artificial:1 choosing:1 exhaustive:1 whose:1 plausible:2 solve:1 drawing:2 ramp:1 statistic:1 advantage:3 sequence:14 essa:10 maximal:1 achieve:1 representational:1 sgi:1 darrell:6 object:12 derive:2 develop:1 coupling:3 pose:3 exemplar:1 strong:1 implemented:1 centered:2 human:1 programmer:1 ultra:1 mapping:2 major:1 estimation:1 proc:2 applicable:1 successfully:1 weighted:1 sidestepped:1 mit:5 sensor:1 rather:1 knob:1 derived:2 focus:1 june:1 indicates:1 sense:2 detect:1 rigid:2 integrated:1 typically:1 entire:2 initially:1 reproduce:1 issue:1 classification:1 priori:1 art:1 special:2 field:2 construct:2 cartoon:2 sampling:2 manually:1 look:1 mimic:2 rosen:1 others:1 few:1 interpolate:2 phase:1 consisting:3 attempt:1 interest:1 capable:1 poggio:5 facial:46 decoupled:1 desired:1 deformation:4 modeling:10 neutral:3 subset:1 successful:2 graphic:5 characterize:1 stored:3 connect:3 adler:1 combined:1 person:1 physic:1 synthesis:8 quickly:2 connecting:1 thesis:1 recorded:1 containing:1 opposed:1 literal:1 conf:1 potential:1 performed:3 view:47 extrapolation:1 lab:5 closed:3 observing:1 characterizes:2 shashua:1 capability:1 waving:1 ensemble:4 correspond:1 spaced:1 yield:2 blink:2 plastic:1 drive:2 straight:1 tissue:3 trevor:5 colleague:1 involved:1 workstation:1 gain:1 proved:2 dimensionality:2 response:4 done:2 stage:2 correlation:16 hand:6 assessment:1 quality:1 ieice:1 effect:1 normalized:4 contain:1 brunelli:2 during:1 criterion:3 nonrigid:1 stress:1 motion:2 interface:2 passive:2 dedicated:1 image:21 wise:1 novel:1 recently:2 rotation:3 physical:2 tracked:2 overview:1 endpoint:1 attached:1 extend:1 interpretation:1 badler:2 interpret:1 measurement:4 significant:1 ai:3 outlined:1 similarly:1 closing:2 reliability:1 similarity:1 closest:1 own:1 driven:7 yi:1 muscle:18 greater:1 paradigm:6 signal:1 ii:2 neurally:4 desirable:2 full:9 multiple:1 match:7 faster:1 gesture:5 offer:1 long:1 paired:1 controlled:1 prediction:1 j3:1 assesment:1 vision:6 physically:5 represent:1 achieved:1 animating:2 addition:1 sends:1 hz:1 acheive:1 flow:6 smile:4 extracting:1 intermediate:2 enough:1 rendering:2 expression:46 o0:1 render:1 action:3 useful:3 detailed:1 offacial:1 processed:2 generate:3 canonical:2 estimated:1 overly:1 track:2 anatomical:2 wireframe:1 controled:1 promise:1 group:1 threshold:2 authentic:1 interpolator:3 rectangle:1 graph:1 geometrically:1 sum:1 run:2 inverse:1 powerful:1 topologically:1 correspondence:1 occur:2 optic:2 constraint:5 scanned:1 alex:5 afforded:1 conferencing:2 scene:1 personalized:1 interpolated:4 aspect:3 speed:1 span:1 performing:1 optical:3 rendered:5 marking:2 creative:1 across:4 character:2 making:1 anatomically:3 invariant:1 visualization:1 previously:1 describing:1 mechanism:2 needed:1 available:1 apply:2 actuator:1 generic:1 indirectly:1 appropriate:1 top:1 clustering:2 include:1 emotional:1 restrictive:1 especially:1 establish:1 warping:1 skin:4 implied:1 added:1 surgery:1 strategy:3 traditional:1 unable:1 separate:1 simulated:1 evenly:1 topic:1 water:4 modeled:1 difficult:3 implementation:3 reliably:1 perform:1 allowing:1 finite:1 tele:1 pentland:9 extended:2 frame:7 intensity:1 pair:2 required:1 specified:1 connection:2 nip:1 trans:1 able:2 usually:2 pattern:2 perception:3 eyebrow:2 reliable:1 including:1 video:9 mouth:3 power:1 critical:1 natural:1 rely:1 hybrid:1 force:2 scheme:1 eye:12 picture:2 created:1 mase:2 extract:1 coupled:1 geometric:2 fully:1 generation:1 limitation:1 accelerator:1 degree:1 conveyed:1 share:2 translation:1 infeasible:1 terzopoulos:2 template:3 face:42 eyelid:1 distributed:6 dimension:2 contour:1 forward:1 replicated:3 transaction:1 xi:2 search:1 nature:1 robust:5 irfan:5 investigated:1 complex:3 interpolating:1 domain:7 constituted:1 animation:8 allowed:2 augmented:1 board:1 fashion:1 position:1 explicit:2 wish:1 exercise:1 lie:1 perceptual:12 hislher:1 dozen:1 explored:2 ci:1 illustrates:1 surprise:3 simply:1 appearance:1 beymer:3 contained:1 tracking:9 acquiring:1 acm:2 rbf:9 change:1 determined:1 decouple:1 invariance:1 mark:1 actuation:3 evaluate:1 |