Anda di halaman 1dari 10

A New Method For Automatic Signature Recognition

Using Artificial Neural Network


Shah Mohammad Walid1
1 Dept. Of ECET,Minnesota State University-Mankato
shah.walid@mnsu.edu

Abstract. Artificial neural networks (ANNs) have been recognized as universal


estimators and have become very useful in the field of pattern recognition. This
paper refers to a signature recognition algorithm based on a new learning
technique, gradient descent with adaptive learning rate and momentum
backpropagation technique. The proposed technique of learning improves
backpropagation algorithm with better accuracy and speed up the entire
learning process. The performance is analyzed with different activation
functions and learning techniques.

Keywords: Artificial neural network, backpropagation, gradient descent with


adaptive learning and momentum backpropagation, signature recognition,

1 Introduction

The two fundamental aspects of neural network, first it’s massively parallel
distributed structure and second its generalization of learning, have given it the
capability of being a universal approximator. Neural network also have the property
of being adaptive, which makes them very powerful in application as well as in
signature recognition. The training algorithm used has an effect on issues such as
learning speed, stability and weight convergence. In this paper, a technique for
improving the learning speed is proposed. Improvisation of learning speed would give
a new dimension to the application of neural network in the field of off-line pattern
recognition.
There are several implementations for signature recognition. Justino, Bortolozzi
and Sabourin proposed an off-line signature verification system using Hidden Markov
Model [1]. The Hidden Markov Model consists of a number of states, the transitions
between which depend on the occurrence of some symbol. Hence, the system is not
totally prepared to discriminate small differences between the genuine model and a
test sample. It obviously depends upon the probability distribution. Zhang, Fu and
Yan (1998) proposed handwritten signature verification system based on Neural ‘Gas’
based Vector Quantization [2]. In this paper signature recognition is discussed with a
special vector quantization method. Vélez, Sánchez and Moreno proposed robust off-
line signature verification system using compression networks and positional
cuttings[3]. Arif and Vincent (2003) concerned data fusion and its methods for an off-
line signature verification problem which are Dempster-Shafer evidence theory,
Possibility theory and Borda count method [4]. This paper discusses about the
ultimate failure of Borda count method. Chalechale and Mertins used line segment
distribution of sketches for Persian signature recognition [5]. Here horizontal and
vertical line segmentation technique is proposed. Sansone and Vento (2000) increased
performance of signature verification system by a serial three stage multi-expert
system [6]. Another new learning method, Support vector machine is introduced by
V. Vapnik et al. [7], [8]. Support vector machines, are another category of universal
feedforward networks, are proposed for pattern classification and non linear
regression.
Multilayer perceptrons are widely used for pattern classification. Backpropagation
algorithm is the most dominant technique, so far proved, for network training.
Backpropagation algorithm can be speed up by considering the adaptive learning rate
and momentum individually. In this paper we proposed for backpropagation
algorithm with a new technique, which in fact associates both adaptive learning rate
and momentum. And the result is analyzed and the learning rate is found improving.
There are two major types of learning methods-one is supervised learning and
other is unsupervised learning. For signature recognition, both supervised and
unsupervised learning is used. For the training session unsupervised learning is used
whereas for classification supervised network is used. So far, the most efficient
algorithm for training a network is found the backpropagation algorithm. This
algorithm is, actually, based on Error Correction Learning. One of the most
important parameter is rate of learning (η ). The rate of learning can be improved by
incorporating the term, usually a positive number called momentum constant (α ). It
actually stabilizes the performance. Thus the technique of learning improves and
faster response can be found.
The application of our proposed technique in pattern association and classification,
signature recognition as well, however, improvises the performance and accuracy.
First the methodology of application will be discussed in section 1. A simulator,
which has been created by MATLAB, is used to simulate the results of the method.
The recognition process and learning process, both are performed through this
simulator. The recognition accuracy is also determined. In section 3 the result of the
simulator is discussed. Besides, various activation functions and other techniques are
also used for comparison using this simulator. So far, the speed up in learning process
by the proposed technique, with respect to others, is illustrated in section 4. Section 5
and section 6 will discuss about the problems and limitations and the ultimate
achievement of this paper.

2 Methodology

For signature recognition, the neural network acts as both pattern associator and
classifier. To recognize signatures, first the knowledge on signature should be must.
The global features of signatures are as follow. Fig.1 illustrates the global features of
signatures.
• Almost constant Height-Width ratio.
• Constant signature area
• Maximum horizontal and vertical histogram
• Almost fixed horizontal and vertical centre.

(a) (b) (c) (d) (e)

Fig. 1. Features of signatures (a) Preprocessed signature (b) Height (c) maximum vertical
histogram (d) maximum horizontal histogram (e) Horizontal centre and vertical centre.

First, the signature pattern is to be taken. The signatures can be taken as image
from a scanner. It should be mentioned that a particular format (ten different
signatures in a row and five patterns in column) should be maintained. The signatures
are separated creating different individual regions for each and the regions are
labeled. Each signature is resized and reconstructed. The procedure is discussed later
in this chapter.
After image preprocessing the signature is found within a 70x50 matrix which is
later converted to a 7x5 matrix by summing each 10x10 matrix. From the value of
each element of 7x5 matrix three particular parameters area, bounding box and
centroid, can be obtained. Area is the actual number of pixels in the region, whereas
bounding box indicates the smallest rectangle containing the region. The other
parameter centroid indicates the centre of mass of the region. In Fig.2 the bounding
box and centroid are illustrated. Thus the feature is extracted. And these features are
kept into the 35 element vector cell matrix in which every cell contains the
aforementioned three parameters. These boxes are displayed during feature extraction
process.
Centroid

Bounding
box

Fig.2. Illustration of centroid and bounding box. The region consists of the shaded pixels; the
dotted line shows the bounding box and black dot is the centroid.

When the signature, which is to be recognized, is taken as image it not only


performs almost same actions as previous, rather something enhanced. The noise
effect is removed from the sample by using functions like image dilation, fill up etc.
This sample is also made to binary and converted into a 35 element vector matrix and
111111111111001111111111111111110001111111111111
111
feature extraction is held in previous way. 111 Fig. 3 illustrates the sequential image
111111111100000111111111111111100000111111111111

preprocessing. Thus the sample 111111111000110011111111111111000110111111111111


is fed to the 111
neural network for classification. It
should be mentioned here, all these actions111 are performed in MATLAB and its
111111110011110011111111111111001111111111111111

111111100111110011111111111110011111111111111111
associative functions. 111
111111000101100011111111111100011111111111111111
111
111111100101100111111111111100111111111111111111
111
111111100011100011111111111100111111111111110000
000
111111111111000000111111111000000011111100000001
111
111111111110011110011100000000000000010000011111
111
111111111111111111000100010000001000000000111111
111
111111111111111111000100111001001111001100000111
111
111111111111111111000000010001001111111100000111
111
111111111111111111000110010001000000001110000111
111
111111111111111111000110010001000000000100100111
111
111111111111111110001100010001000001000111100111
111
000111111111111100000000110011000111101111100111
111
000000111111110001111111110011001111111111100111
(a) (b) 111
111000000000000111111111111111111111111111100111
(c)
111
111111111000111111111111111111111111111111100111
111
Fig.3. (a) sample input image (b) corresponding binary format (c) feature extracted format.
111111111111111111111111111111111111111111001111
111
111111111111111111111111111111111111111111001111
111
111111111111111111111111111111111111111111001111
111

The neural network, which is created for pattern association and classification, has
some specific criterions. Since 35 element vector matrix is obtained from the feature
extraction process, the 35 element vector matrix is to be fed at network input. Thus
the network contains 35 input nodes and the input data is fed at these nodes paralelly.
Two hidden layers, with logistic sigmoidal activation functions in both, are connected
between the input and output layer. Tan hyperbolic and linear functions are also used
only for comparison purpose. The weight matrices and biases are taken randomly.
The output layer contains 10 output nodes since ten different signatures are to
classify. Gradient descent with adaptive learning rate and momentum
backpropagation technique is used as learning algorithm, which is discussed earlier.
Also, other techniques are used but those are for comparison purpose. Momentum
constant is first kept at 0.95 and the maximum number of epoch is kept to 5000. Sum
squared error (SSE) function is used as performance function. The performance goal
is kept to 0.1. The four stopping criterion are that user stop, minimum gradient reach,
maximum epoch reach, minimum performance goal meet. If any one of these four
criterions occurs, the training session dies out. At classification stage, from output
layer ten outputs are found which are stacked in a vector matrix and the maximum
value among the matrix elements indicates the corresponding signer.
Another important thing is that, a text file containing the name or database of the
signers should be attached with the image file to be loaded. The text file should have
the same title as the image file. The sequence of information should be maintained.
This text file is used to show the output.
The entire procedure can be briefly illustrated with a flow chart shown in Fig.4.
Here the major blocks are shown. In this particular case, the neural network performs
two functions. First, it gets trained with input dataset and then it classifies the pattern
with trained and sample dataset. It should be mentioned here that the sample dataset
also needs to be preprocessed.

Start

Taking input
dataset

Image Preprocessing

Image dilation and


reconstruction

Feature Extraction

Neural
Dataset For
Network
Test

Classified
Output

End

Fig.4. Flow chart of signature recognition system

3 Results

Automatic signature recognition system, with our proposed technique, recognizes


signatures quite accurately. Two tests were performed. One is for incomplete training
and other is for complete training. Two test results were found and tabulated in
Table1. It should be mentioned here that the accuracy of the two tests is quite closer.
This is because the training performance is found almost near to the predefined goal
(performance was 1.003547) for incomplete training. For the test, each signature is
taken 10 times and tested by the simulator.

Table 1.Accuracy test result

Sample pattern Test1 Test2


Signature 1 100% 100%
Signature 2 100% 100%
Signature 3 90% 100%
Signature 4 90% 90%
Signature 5 100% 100%
Signature 6 60% 80%
Signature 7 80% 100%
Signature 8 80% 90%
Signature 9 90% 100%
Signature 10 100% 100%
Average 89% 96%

Besides testing the accuracy, the most suitable activation function is obtained.
Activation functions are used in the output of two hidden layers. For a particular
learning technique different activation functions like logistic sigmoidal, tan
hyperbolic and pure linear functions are used. Among those logistic sigmoidal
functions proved most dominant. The comparison is illustrated in the next section.

4 Comparison

The simulator is compatible for different learning techniques. Moreover, the


activation functions can also be changed and analyzed. On the other hand the
flexibility of variations in maximum number of epochs, momentum constant and goal
give a new dimension in performance analysis of the signature recognition system.
For different learning technique the different result is found. The result is tabulated in
table 2. Here the number of epochs comes to consideration. It shows the acceleration
of the learning procedure.

Table 2.Comparison of different techniques of learning

Gradient descent with Gradient descent with Gradient descent with


adaptive learning rate momentum adaptive learning rate
backpropagation backpropagation and momentum
technique technique backpropagation
technique
Number Performance Number Performance Number Performance
of epochs of of
epochs epochs
Test 1 260 0.0984638 5000 1.00675 229 0.0787023
Test 2 334 0.988488 5000 0.99529 212 0.0804202
Test 3 291 0.0994702 5000 5.39558 234 0.0756297
Test 4 257 0.0976441 5000 12.6805 229 0.0787023
Test 5 356 0.0988924 5000 0.85535 191 0.0895776

Fig. 5 shows the variations in learning for different learning techniques. The result
of first test is illustrated in this figure. It should be mentioned here that two of these
three methods can meet the goal within maximum assigned epochs and the rest one
cannot. From Fig. 5 it can be said that gradient descent with adaptive learning rate
and momentum backpropagation technique is faster than that of other, alternatively,
less epochs are required.

Comparing these three techniques, applied for the signature recognition system, it
is found that our proposed technique is better and faster than other techniques with a
reasonable accuracy.

Fig. 5. Comparison of different learning techniques. Here tech1 stands for gradient descent
with adaptive learning rate back propagation technique; tech2 stands for gradient descent with
adaptive learning rate and momentum backpropagation technique; tech3 stands for gradient
descent with momentum backpropagation technique.

For different activation function, taking the gradient descent with adaptive learning
rate and momentum backpropagation technique, different characteristics of learning is
obtained. Fig. 6 illustrates the different criteria for different activation functions.
Figure 4.2 Comparison of different activation functions. Here for two of the activation
functions the goal is not met. Only for log sigmoid function the performance is converged to
the goal at 229 epochs.

From the above point of view, it is quite clear that logistic sigmoidal function is the
most suitable for training.

Along with this, another test was performed. For different types of training method
the classification error was obtained. The test result is tabulated in table 3. It should
be mentioned here that for two cases the performance goal was met. Only for gradient
descent with momentum backpropagation technique the performance goal did not
meet conveying a value of 6.69. Five patterns are used here for testing.

Table3. Classification error

Classification error in percentage


Gradient descent Gradient descent Gradient descent
Sample pattern with adaptive with momentum with adaptive
learning rate backpropagation learning rate and
backpropagation technique momentum
technique backpropagation
technique
Signature1 0 100 0
Signature2 20 0 0
Signature3 0 40 0
Signature4 20 0 20
Signature5 40 80 20
Signature6 60 100 0
Signature7 40 80 40
Signature8 0 40 0
Signature9 20 60 0
Signature10 60 100 0
Average 26 60 8

5 Discussion

Since the initial weight and biases are taken as random number, the learning
characteristic is different at different times. They are not exactly equal, but lies within
a specific range. This is why, sometimes it become difficult to separate different close
results. The major advantage of this system is it can also recognizes characters, if it is
being trained so. Moreover, the system is flexible with respect to network parameters
and functions. The remarkable accuracy and noise removal feature gives it a new
dimension.
The system is limited for discontinuous signatures. The system cannot recognize
any discrete signature. This is due to the input taking format. If single signature is
taken as input, this problem can be overcome, but the system would be time
consuming and more complex for user. Another problem is that, a text file containing
the names of the signers is to be loaded with the image. Otherwise it cannot work.

6 Conclusion

The analysis shows that our proposed technique is more reliable and fast
responding than that of the other techniques previously used. The dependency of
learning on activation functions are discovered one more time. The suitable activation
function is found for pattern recognition. The accuracy can be improved more if the
feature extraction is performed by more data handling, though it would be more
complex and time consuming process.

References

1. E. J. R. Justino, F. Bortolozzi and R. Sabourin : Off-line Signature Verification


Using HMM for Random, Simple and Skilled Forgeries. ICDAR 2001,
International Conference on Document Analysis and Recognition, vol. 1, pp.
105--110. 2001
2. B. Zhang, M. Fu and H. Yan : Handwritten Signature Verification based on
Neural ‘Gas’ Based Vector Quantization. IEEE International Joint
Conference on Neural Networks, pp. 1862-1864, May 1998.
3. J. F. Vélez, Á. Sánchez , and A. B. Moreno: Robust Off-Line Signature
Verification Using Compression Networks And Positional Cuttings. Proc. 2003
IEEE Workshop on Neural Networks for Signal Processing, vol. 1, pp. 627-636,
2003.
4. M. Arif and N. Vincent: Comparison of Three Data Fusion Methods for an
Off-Line Signature Verification Problem. Laboratoire d’Informatique,
Université de François Rabelais, 2003.
5. 5.A.Chalechale and A. Mertins: Line Segment Distribution of Sketches for
Persian Signature Recognition. IEEE Proc. TENCON, vol. 1, pp. 11–15, Oct.
2003
6. Sansone and Vento: Signature Verification: Increasing Performance by a
Multi-Stage System. Pattern Analysis & Applications, vol. 3, pp. 169–181,
2000.
7. C.Cortes and V.Vapnik: Support-vector networks. Machine Learning. vol.
20, pp:273-297, Nov. 1995.
8. V.N.Vapnik: The Nature of Statistical Learning Theory. Springer, 1995.
9. Joachims, T: Text categorization with support vector machines: Learning
with many relevant features. Proceedings of the Tenth European Conference
on Machine Learning. Berlin:Springer, pp. 137 -142, 1998.
10. H.Baltzakis and N. Papamarkos: A new signature verification technique
based on a two-stage neural network classifier”, Pergamon, pp 95-103, 2001.
11. H. Gómez-moreno, P. Gil-jiménez, S. Lafuente-arroyo, R. Vicen-bueno and
R. Sánchez-montero: Color images segmentation using the Support Vector
Machines.
12. S.Haykin: Neural networks: A comprehensive foundation.
13. N.K.Bose, P.Liang: Neural network fundamentals with graphs, algorithms
and applications.

Anda mungkin juga menyukai