Handwritten Character Recognition Using Feed-Forward Neural Network Models

download Handwritten Character Recognition Using Feed-Forward Neural Network Models

of 10

Transcript of Handwritten Character Recognition Using Feed-Forward Neural Network Models

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    1/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    54

    HANDWRITTEN CHARACTER RECOGNITION USINGFEED-FORWARD NEURAL NETWORK MODELS

    Nilay Karade 1, Dr. Manu Pratap Singh 2, Dr. Pradeep K. Butey 3

    1A-304, Shivpriya Towers, Jaitala, Nagpur-440039, Maharshatra, India2Department of Computer Science, Dr. B.R.Ambedkar University, Khandari,

    Agra - 282002, Uttar Pradesh, India3HOD (Computer Science), Kamala Nehru Mahavidyalaya, Nagpur, India

    ABSTRACT

    Handwritten character recognition has been vigorous and tough task in the field of patternrecognition. Considering its application to various fields, a lot of work is done and is beingcontinuing to improve the results through various methods. In this paper we have proposed a systemfor individual handwritten character recognition using multilayer feed-forward neural networks. Forthe experimental purpose we have taken 15 samples of lower & upper case handwritten Englishalphabets in scanned image format i.e. 780 different handwritten character samples. There are twomethods of feature extraction are used to construct the pattern vectors for training set. This trainingset is presented to the six different feed-forward neural networks namely newff, newfit, newpr,

    newgrnn, newrb and newrbe . The test pattern set is used to evaluate the performance of these neuralnetworks models. The results are compared to find the accuracy in recognition of the respectivemodels. The number of hidden layer, number of neurons in hidden layer, validation checks andgradient factors of the neural networks models are taken into consideration during the training.

    Keywords: Character Recognition, multilayer feed-forward Artificial Neural Network,Backpropagation, Handwriting recognition, Pattern Classification

    1. INTRODUCTION

    These days computer have been penetrated in every field and the work is being done at a

    higher speed with greater accuracy. Pattern recognition through computer is a challenging task andthis task becomes more critical if the pattern is in the form of handwritten curve script. Pattern

    INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING &TECHNOLOGY (IJCET)

    ISSN 0976 – 6367(Print)ISSN 0976 – 6375(Online)Volume 6, Issue 2, February (2015), pp. 54-74© IAEME: www.iaeme.com/IJCET.aspJournal Impact Factor (2015): 8.9958 (Calculated by GISI)www.jifactor.com

    IJCET

    © I A E M E

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    2/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    55

    recognition, as a subject, spans a number of scientific disciplines, uniting them in search for asolution to the common problem of recognizing the pattern of a given class and assigning the nameof identified class. Pattern recognition is the categorization of input data into identifiable classesthrough the extraction of significant attributes of the data from irrelevant background detail. Apattern class is a category determined by some common attributes. It is true that the olderhandwritten documents are digitized but the 100% automation of work cannot be achieved. Thehandwriting recognition has helped a lot to the advancement of automation process [1]. Thehandwriting recognition systems are broadly classified into two types, namely online and offlinehandwritten recognition. In online approach, the two dimensional coordinates of the consecutivepoints are symbolize as a function of time. Also the sequence of the strokes made by the writer is onhand. Whereas in the case of off-line handwriting recognition approach the written script is capturedwith the help of devices like scanner and the whole script is available as an image [2]. When boththese approaches are compared, it has been found that due the temporal information available withthe online approach, it is superior to that of off line approach [3]. On the other hand in the off-linesystems, the neural networks have been productively used for capitulate comparably high recognitionaccuracy levels [1]. A number of applications such as document analysis, mailing addressinterpretation, bank processing etc. require offline handwriting recognition system [1, 4]. Thus, theoff-line handwriting recognition enjoys the first choice by many researchers in order to investigateand discover the novel methods that would get better recognition correctness. It is widely used inimage processing, pattern recognition, and artificial intelligence.

    During the last few years the researchers have proposed many mathematical approaches tosolve the pattern recognition problems. Recognition strategies heavily depend on the nature of thedata to be recognized. In the cursive case, the problem is made complex by the fact that the writing isfundamentally ambiguous as the letters in the word are generally linked together, poorly written andmay even be missing. On the contrary, hand printed word recognition is more related to printed word

    recognition, the individual letters composing the word being usually much easier to isolate and toidentify. As a consequence of this, methods working on a letter basis (i.e., based on charactersegmentation and recognition) are well suited to hand printed word recognition while cursive scriptsrequire more specific and/or sophisticated techniques. Inherent ambiguity must then be compensatedby the use of contextual information.

    Neural network computing has been expected to play a significant role in a computer-basedsystem of recognizing handwritten characters. This is because a neural network can be trained quitereadily to recognize several instances of a written letter or word, and can then be generalized torecognize other different instances of that same letter or word. This capability is vital to therealization of robust recognition of handwritten characters or scripts, since characters are rarelywritten twice in exactly the same form. There have been reports of successful use of neural networks

    for the recognition of handwritten characters [11, 12], but we are not aware of any generalinvestigation which might shed light on the systematic approach of a complete neural networksystem for the automatic recognition of cursive character. The techniques of artificial neuralnetworks are widely used for pattern recognition task over the conventional approaches to handlesuch type of problem due to the following reasons:1. The same alphabet character written by the same person can vary in shape, size and style.2. It is not only the case with same person but also the shape, size and style of the same character

    can vary from person to person.3. Character image scanned in offline method might have poor quality due to noise present within

    it.4. As there are no pre defined rules about the look of the visual character, the rules should be

    heuristically deduced form the set of sample data. The human brain by its very nature does thesame thing using the features discussed in the following two points.

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    3/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    56

    5. The human brain can read handwritings of various people having different fashion of writingbecause it is adaptive to slight variations and errors in pattern.

    6. It can take hold of new styles present in the character due to its ability of learning fromexperiences with no time.J. Praddep, E.Srinivasan,and S.Himavathi [1] have proposed a handwritten character

    recognition system using neural network by means of diagonal based feature extraction method.They have stated with the binarization of the image which results in binary image, which furtherundergoes the edge detection and dilation and then segmentation. In the process of segmentation aseries of characters is decomposed into sub image of each individual character, each of which isconverted into 90x60 pixels for classification and recognition process. Each character imageobtained in such a way that it is divided into 54 equal zones, each of size 10 x10 pixels andthen features are extracted from each zone pixels by moving along its diagonals. They haveended up with 54 features for each of the character. Another feature extraction method gives them 69features by averaging the values placed in zones row wise and column wise. A feed forward backpropagation neural network having two hidden layers with architecture of 54-100-100-38 is used toperform the classification with both the features with vertical, horizontal and diagonal orientationand have found 92.69, 93.68 , 97.80 percent accuracy and 92.69, 94.73, 98.54 percent accuracy,respectively.

    Kauleshwar Prasad, Devvrat C. Nigam, Ashmika Lakhotiya and Dheeren Umre [3] haveconverted the character image into a binary image, and then apply character extraction algorithm inwhich it has empty traverse list initially. A row is scan pixel by pixel and on getting black pixel, it ischecked if it is already in the traverse list. It is checked that if it is already there then it is ignored,otherwise added to the traverse list using edge detection algorithm. They have claimed to have goodresults by using feed-forward Backpropagation neural network and also stated that poorly chosenfeature extraction method gives poor results.

    Ankit Sharma and Dipti R Chaudhary [4] have achieved the accuracy of 85%, using feedforward neural network. The special form of Reduction is used which includes the noise removal andedge detection for the feature extraction of grayscale images.

    Chirag I Patel, Ripal Patel, Palak Patel [5] have achieved the accuracy of 91%, 889%, 91%,91%, 94%, 94% using different models of Backpropagation neural networks. After characterextraction and edge detection from the document, it goes under the process of normalization wherethe images having various sizes are normalized to a uniform size. The resultant image is applied with‘Line Fitting’, a skew detection technique to correct this skewedness, in which it is rotated by anangle θ . The constructed pattern from this method is further used for the training by Backpropagationalgorithm of feed-forward multilayer neural networks.

    Anita Pal and Dayashankar [7] have used a multilayer perceptron with one hidden layer to

    recognize Handwritten English Character. Boundary tracing along with Fourier Descriptor is used toextract the feature from the handwritten character. By analyzing its shape and judge against itsfeatures, a character is identified. Test result was found to have fine recognition accuracy of 94%for handwritten English characters with less training time.

    The genetic algorithm is used with feed forward neural network architecture as the hybridevolutionary algorithm [27] for the recognition of handwritten English alphabets. In this paper eachcharacter is considered as the gray level image and divided into sixteen parts. The mean of each partis considered as one feature of the pattern. Thus, there are sixteen features in real numbers are usedas the pattern vector for each image. The trained network performed well for the patternclassification for test patterns.

    In this paper we consider the two approaches of feature extraction from the images of

    handwritten capital and small letters of English alphabets. The first method of feature extraction usesthe row wise mean value of the pixels for a processed image of size n x n. The second method

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    4/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    57

    consider the each pixel value of the dilated image of size n x n. These features are used to constructthe pattern vectors. The two training sets are formed from these samples examples of pattern vectors.There are six different feed forward neural networks models are used with six different learningmethods. The performances of these neural networks with different learning rules are analyzed. Therate of recognition for patterns from the test pattern set is also evaluated. The performance evaluationindicates that the Radial bias function (RBF) neural network architecture performs better than otherneural network models for both the methods of feature extraction. The rate of recognition for testpattern set in RBF is found better with respect to other neural network models.

    Rest of the paper is containing 6 sections. Section 2 of the paper describes the featureextraction methods for handwritten English characters. The section 3 discusses about feed forwardneural networks and Backpropagation learning and Radial basis function. The section 4 describes theexperiment and simulation design. The Section 5 presents the simulated results and discussion.Section 6 describes the conclusion followed by the references.

    2. FEATURE EXTRACTION

    Feature extraction and selection can be defined as extracting the most representativeinformation from the raw data, which minimizes within class. The pattern variability whileenhancing between class pattern variability so that, a set of feature are extracted from each class thathelps to distinguish it from other classes, while remaining invariant to characteristic differenceswithin the class. Here we are considering the feature extraction from the input stimuli with twomethods namely the row wise mean of pixel from a scanned image and each pixel value of theimage. In our approaches we have considered the input data in the form of fifteen different set ofeach hand written capital and small English characters by five different peoples. It is quite naturalthat the five different people considered the different hand writing and different writing style for

    every character. So, in this way we have total 780 samples. Among these 780 samples we used 520samples for training and the remaining 260 samples were used in test pattern set. Now to prepare ourtraining set of input output pattern pairs, we consider each scanned hand written character as a colorbit map image. This color bitmap image of a character is now changed into gray level image and theninto binary image as shown in figure 1.

    Fig 1 (a ) gray level image Fig 1 (b) Binary Image

    Now we obtain the images after the edge detection and dilation for both the methods offeature extraction. The edged and dilated images can show in figure 2.

    Fig 2 (a) edged image Fig 2 (b) dilate image

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    5/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    58

    Hence to obtain the uniform pattern vector for every input stimulus we make the dilatedimages of equal sizes by resizing the images into the size of 30 x 30 as shown in figure 3.

    Fig 3: uniform resize images

    Now in the first method of feature extraction we construct the pattern vector for the processedimages of English alphabets by taking row wise mean of image of size 30 x 30. The obtained patternvector will represent in column matrix of order 30 x 1. Thus we have an input pattern matrix of order30 x 520 with target output pattern matrix of order 6 x 520 because to distinguish characters fromeach other we require 52 different classes so that we use the 6 binary digits to present the targetoutput pattern vector.

    In the second method of feature extraction we construct the pattern vector for the processedimages of English alphabets by taking each pixel value of the image. Therefore we have an inputpattern vector of size 900 x 1. Thus we have an input pattern matrix of order 900 x 520 with targetoutput pattern matrix of 6 x 520.

    Thus, we have constructed the training set of input output patterns pairs to analyze theperformance of multilayer feed forward neural networks with six different learning methods. Wehave also constructed our test pattern set to verify the performance of networks. Our test pattern setconsist with another set of hand written characters i.e. order 30 x 30 and 900 x 30 for both the

    methods of pattern presentation respectively. Input pattern for these test character set are constructedin same manner as we did for training set pattern.

    3. FEED FORWARD NEURAL NETWORKS MODEL

    The neural approach applies biological concepts to machines for pattern recognition. Theoutcome of this effort is invention of artificial neural networks. Neural networks can be viewed asmassively parallel computing systems consisting of an extremely large number of simple processorswith many interconnections. Neural network models attempt to use some organizational principles(such as learning, generalization, adaptively, fault tolerance, distributed representation, andcomputation) in a network of weighted directed graphs in which the nodes are artificial neurons and

    directed edges (with weights) are connections between neuron outputs and neuron inputs. The maincharacteristics of neural networks are that they have the ability to learn complex nonlinear input-output relationships, use sequential training procedures, and adapt themselves to the data. The mostcommonly used family of neural networks for pattern classification tasks [13] is the feed-forwardnetwork, which includes multilayer perceptron and Radial-Basis Function (RBF) networks. Thesenetworks are organized into layers and have unidirectional connections between the layers. Thelearning process involves updating network architecture and connection weights so that a networkcan efficiently perform a specific pattern recognition task. The increasing popularity of neuralnetwork models to solve pattern recognition problems has been primarily due to their seemingly lowdependence on domain-specific knowledge (relative to model-based and rule-based approaches) anddue to the availability of efficient learning algorithms. Neural networks provide a new suite of

    nonlinear algorithms for feature extraction (using hidden layers) and classification (e.g., multilayerperceptron). In spite of the seemingly different underlying principles, most of the well known neural

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    6/21

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    7/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    60

    where,

    j

    m

    j j xwv ∑

    =

    =0

    (3.3)

    owb = The output at every node can finally calculates by using sigmoid function

    Kxe x f y −+

    ==1

    1)( ;where K is the adaption constant (3.4)

    Figure 4: The Functioning of neural network architecture.

    The supervised learning mechanism is commonly used to train the feed forward multilayerneural network architecture. In this learning process a pattern is presented at the input layer. Thepattern will be transformed in its passage through the layers (hidden) of the network until it reachesthe output layer. The units in the output layer all belong to a different category. The outputs of thenetwork as they are now compared with the outputs as they ideally would have been if this patternwere correctly classified, in the later case unit with the correct category would have had the largestoutput value and the output values of the other output units would have been very small. On thebasis of this comparison all the connection weights are modified a little bit to guarantee that, thenext time this same pattern is presented at the inputs, the value of the output unit that correspondswith the correct category is a little bit higher than it is now and that, at same time, the output valuesof all the other incorrect outputs are little bit lower than they are now. The differences between the

    actual outputs and the idealized outputs are propagated back from the top layer to lower layers tobe used at these layers to modify connection weights. Thus it is consider as back propagationlearning algorithm.

    The Backpropagation (BP) learning algorithm is currently the most popular supervisedlearning rule for performing pattern classification tasks [20]. It is not only used to train feed forwardneural networks such as the multilayer perceptron, it has also been adapted to recurring neuralnetworks [21]. The BP algorithm is a generalization of the delta rule, known as the least mean squarealgorithm. Thus, it is also called the generalized delta rule . The BP overcomes the limitations of theperceptron learning enumerated by Minsky and Papert [22]. Due to the BP algorithm, the MLP canbe extended to many layers. This algorithm propagates backward the error between the desired signaland the network output through the network. After providing an input pattern, the output of thenetwork is then compared with a given target pattern and the error of each output unit calculated.This error signal is propagated backward, and a closed-loop control system is thus established. The

    Weights

    SummingFunction

    Bias

    b Activation

    Function

    Local

    Field

    v

    Output

    y

    x1

    x2

    xm

    ω 2

    ∑ )(−ϕ

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    8/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    61

    weights can be adjusted by a gradient-descent approach. In order to implement the BP algorithm, acontinuous, nonlinear, monotonically increasing, differentiable activation function is required asLogistic Sigmoid function or hyperbolic tangent function.

    So that, to provide the training for multi-layer feed forward network to approximate anunknown function, based on some training data consisting of pairs S z x ∈, , the input pattern vector

    x represents a pattern of input to the network, with desired output pattern vector z from the trainingset S. The objective function for optimization or minimization is defined in the sum ofinstantaneously squared error as:

    ∑=

    −= J

    j j j

    P S T E 1

    2)(21 (3.5)

    where 2)( j j S T − is the squared difference between the actual output of the network on the outputlayer for the presented input pattern P and the target output pattern vector for the pattern P. All the

    network parameters ( )1−mW and mθ , m = 2 M, can be combined and represented by the matrix

    ijwW = . So that, the error function E can be minimized by applying the gradient-descent procedure

    as:

    W E

    W ∂∂

    −=∆ η (3.6)

    where η is a learning rate or step size, provided that it is a sufficiently small positive number.Applying the chain rule the equation (3.6) can express as

    ( ) ( )

    ( )

    ( )mij

    m j

    m j

    mij w

    uu E

    w E

    ∂∂

    ∂∂=

    ∂∂

    +

    +

    1

    1 (3.7)

    while( )

    ( ) ( )( ) ( ) ( )( ) ( )mim jmm jm

    ijm

    ij

    m j

    oowww

    u=+

    ∂=

    ∂∑ +

    +1

    1

    θ (3.8)

    and( ) ( )

    ( )

    ( ) ( )( ) ( )( )11

    1

    1

    11++

    ++

    +

    ++ ∂∂

    =∂

    ∂∂

    =∂

    ∂m

    jm

    jm j

    im j

    m j

    m j

    m j

    uo

    E u

    o

    o E

    u

    E φ (3.9)

    For the output unit m=M-1( ) jm j

    eo

    E =

    ∂+1

    (3.10)

    For the hidden units, m = 1 ,2,3………,M − 2,

    ( )1

    221

    2+

    =++ ∑

    +

    ∂=

    ∂ m j

    j

    mm j

    m

    u

    E

    o

    E ω

    ω ω

    ω (3.11)

    Define the delta function by( )

    ( )m p

    m j

    u

    E

    ∂=δ (3.12)

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    9/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    62

    for m = m = 2,3………,M . By substituting (3.7), (3.11), and (3.12) into (3.9), we finally obtain thefollowing equations:For the output units, m = M − 1,

    ( ) ( ) ( ) M j M j j M j ue φ δ −= (3.13)

    For hidden units, m = 1,……..,M − 2,

    ( ) ( ) ( )( ) ( ) 11

    22

    +

    =

    +∑+

    −= m J

    m M j

    M j j

    M j

    m

    ue ω ω

    ω ω δ φ δ (3.14)

    Equations (3.13) and (3.14) provide a recursive method to solve ( )1+m jδ for the whole network. Thus,W can be adjusted by

    ( )( ) ( )m

    im

    jmij

    o E 1+−=

    ∂δ

    ω (3.15)

    For the activation transfer functions, we have the following relations for the logistic function

    ( ) ( ) ( )[ ]uuu φ βφ φ −= 1 (3.16)

    For the tanh function

    ( ) ( )uu2

    1 φ β φ −= (3.17)

    The update for the biases can be done in two ways. The biases in the ( m+1) th layer i.e. θ (m+1) can be expressed as the expansion of the weight W (m), that is, ( ) ( ) ( )m J

    mmm 1,01,0

    1 ..........,.........+

    =+ ω ω θ .

    Accordingly, the output o(m) is expanded into ( ) ( ) ( )m J mm

    mooo ....,,.........,1 1= . Another way is to use a

    gradient-descent method with regard to θ (m), by following the above procedure. Since the biases canbe treated as special weights, these are usually omitted in practical applications. The algorithm is

    convergent in the mean ifmax

    20

    λ η

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    10/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    63

    Whereas the weight update between hidden layer and input layer can be represent as:

    ( ) ( )( )( )sw

    sww E

    swho

    N

    iih

    ihih ∆−

    +∆+∂∂

    −=+∆ ∑= α

    α η 1

    11

    1

    (3.19)

    Where α is the momentum factor, usually 0 < α ≤ 1.The BP algorithm is a supervised gradient-descent technique, wherein the MSE between the

    actual output of the network and the desired output is minimized. It is prone to local minima in thecost function. The performance can be improved and the occurrence of local minima reduced byallowing extra hidden units, lowering the gain term, and with modified training with different initialrandom weights. These are also efficient variant of Backpropagation learning algorithms likeconjugate descent, Levenberg-Marquardt Backpropagation and Radial bias functions. There aredifferent six neural networks are used with these learning techniques namely feed forward network,fitting network, pattern recognition, generalized regression neural network and Radial basis neuralnetworks. These models and learning algorithms are used to improve the performance of feedforward multilayer network architecture for the given training set.

    3.1 Radial Basis FunctionIn this section, we investigate the network structure related to the multi layer feed-forward neural

    network (FFNN), implemented using the Radial Basis Function. RBF networks emulate the behavior ofcertain biological networks. RBF-MLP is essentially feed forward neural network with three layersnamely Input, Hidden and Output. The single hidden layer consists of the locally tuned or locallysensitive units, and the output layer (in most cases) consists of binary responsive units. In the hidden layerunits, the unit response is localized and decreases as a function of the distance of input from the unit’sreceptive field center. The RBF-MLP uses a static Gaussian function as the nonlinearity for the hiddenlayer neurons. The Gaussian function responds only to a small region of the input space where theGaussian is centered. The key to a successful implementation of these networks is to find suitable centersfor the Gaussian functions [25] in supervisory mode. The process starts with the training of input layer. Itsfunction is to obtain the Gaussian centers and the widths from the input samples. Thus achieved centersare then prearranged within the weights of the hidden layer. The output of this layer is derived from theinput samples weighted by a Gaussian combination. The advantage of using the radial basis function isthat it discovers the input to output map using local approximations [26]. Usually the supervised segmentis simply a linear combination of the approximations. Since linear combiners have few weights, thesenetworks train extremely fast and require fewer training samples.

    In contrast to the classical MLP the activation of a neuron is not given by the weighted sum of itsall inputs but by the computation of a RBF. The RBF that we use is Gaussian Function, which can be

    expressed as:

    (3.1.1)

    Where ф is the Gaussian Function, x is the input to the neuron i, µ i is the basis of neuron i and σ i is the amplitude of neuron i. The input layer has i nodes, the hidden and the output layer have k and jneurons, respectively. Each input neuron corresponds to a component of an input vector x. Each node inthe hidden layer uses an RBF as its non linear activation function and performs a non-linear transform ofthe input. The output layer is a linear combiner, mapping the nonlinearity into a new space. The RBF-MLP can achieve a global optimal solution to the adjustable weights in the minimum MSE by using the

    linear optimization method. Therefore for an input pattern x, the output of the jth

    node of output layer canbe defined as:

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    11/21

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    12/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    65

    ∆ ∑ ∗ − ∗ ∗

    (3.1.9)

    And, ∆ ∑ ∗ − ∗ ∗

    (3.1.20)

    We have from equations (3.1.8), (3.1.9) & (3.1.20) the expressions for change in weight vector &Radial basis function parameters to accomplish the learning in supervised way. The setting of the RadialBasis function parameters with supervised learning represents a non linear optimization problem whichwill typically be computationally intensive and may be proven to find local minima of the error function.Thus, for reasonable well localized RBF, an input will generate a significant activation in a small regionand the opportunity of getting stuck at a local minimum is small. Hence, the training of the network for Lpattern pair i.e. (x l, y l) will accomplish in iterative manner with the modification of weight vector.

    4. EXPERIMENT AND SIMULATION DESIGN

    In this paper we have implemented two feature extraction methods on six different artificialneural network models in Matlab, namely feed forward network (newff) , fitting network (newfit) ,generalized regression (newgrnn) , pattern recognition (newpr) , radial basis network (newrb) andexact radial basis network (newrbe) with Levenberg-Marquardt Backpropagation and Radial biasfunctions . In this simulation design for each neural network model we have created 2 networks, onefor lower case another for upper case characters which consume the input retrieved from first featureextraction method. Similarly another two networks are created for the same models of neuralnetworks those use data generated from second method of feature extraction. Thus, there are four

    neural networks created for each model of neural network. The architectural detail of the each modelis presented in table 1, 2, 3, 4, 5 and 6 respectively.

    (1) Newff network with Levenberg-Marquardt learning rule

    Table 1: Architecture detail about Newff Description Network 1 Network 2

    Number of hidden layers 3 2Number neurons in hidden layer 37-23-7 21-11Number neurons in output layer 5 5Number of inputs 30 30Transfer function tansig- tansig- tansig tansig- tansigTraining function trainlm trainlmLearning rate 1.0000e-003 1.0000e-003Max number of epochs 1000 1000Error goal 0 0Number of samples of each alphabet for pattern 10 10Number of samples of each alphabet for training 5 5

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    13/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    66

    (2) Newfit network with Levenberg-Marquardt learning rule

    Table 2: Architecture detail about Newfit Description Network 3 Network 4

    Number of hidden layers 3 2Number neurons in hidden layer 31-17-9 21-11Number neurons in output layer 5 5Number of inputs 30 30Transfer function tansig- tansig- tansig tansig- tansigTraining function trainlm trainlmLearning rate 1.0000e-003 1.0000e-003Max number of epochs 1000 1000Error goal 0 0Number of samples of each alphabet for pattern 10 10

    Number of samples of each alphabet fortraining 5 5

    (3) Newgrnn Network with Radial Basis Function

    Table 3: Architecture detail about Newgrnn Description Network 5

    Number of hidden layers 1Number neurons in hidden layer 260Number neurons in output layer 5Number of inputs 30Number of samples of each alphabet for pattern 10Number of samples of each alphabet for training 5

    (4) NewPR network with Levenberg-Marquardt learning rule

    Table 4: Architecture detail about Newpr Description Network 6

    Number of hidden layers 4Number neurons in hidden layer 41-31-17-7Number neurons in output layer 5

    Number of inputs 30Transfer function tansig- tansig- tansig- tansigTraining function trainscgMax number of epochs 1000Error goal 0Number of samples of each alphabet for pattern 10Number of samples of each alphabet for training 5

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    14/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    67

    (5) Newrebe network with Radial Basis Function

    Table 5: Architecture detail about Newrbe Description Network 7

    Number of hidden layers 1Number neurons in hidden layer 260Number neurons in output layer 5Number of inputs 30Number of samples of each alphabet for pattern 10Number of samples of each alphabet for training 5

    (6) Newrb network with Radial Basis Function

    Table 6: Architecture detail about Newrb

    Description Network 8Number of hidden layers 1Number neurons in hidden layer 260Number neurons in output layer 5Number of inputs 30Number of samples of each alphabet for pattern 10Number of samples of each alphabet for training 5

    Therefore six neural network models are used with eight neural network architectures. Thetwo different supervised learning methods are used i.e. Levenberg-Marquardt learning and RadialBasis function approximation. The simulation results are obtained from all these networks for boththe feature extraction methods.

    5. RESULT AND DISCUSSION

    The simulated results are obtained from both the methods of feature extraction with all the sixmodels of neural networks by using Levenberg-Marquardt Backpropagation learning and RadialBasis approximation. The training set consists with handwritten English capital and small alphabets.The performance of neural network model for training and testing is presented with regression valueand regression line for the simulated output values of the neural network models. The performanceof all the six neural network models for training and testing is presented in table 7, 8, 9, 10, 11 & 12and figure 5, 6, 7, 8, 9, 10, 11 & 12.

    Table 7: Simulated Results for Newff model with Levenberg-Marquardt learning rule

    Description Pattern data trainingregression valueAverage of regression value for test

    data samplesNetwork 1 using Feature

    Extraction method 1 0.33743 0.211826

    Network 1 using FeatureExtraction method 2 0.562268 0.201676

    Network 2 using FeatureExtraction method 1

    0.50037 0.20738

    Network 2 using FeatureExtraction method 2 0.24005 0.000335

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    15/21

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    16/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    69

    Figure 8: Performance of Network4 for both the feature extraction methods

    Table 9: Simulated Results for Newgrnn model with Radial Basis Function Approximation

    Description Pattern data training regressionvalueAverage of regression value test data

    samples

    Network 5 using FeatureExtraction method 1 0.556283 0.408253

    Network 5 using FeatureExtraction method 2 1 0.72463

    Figure 9: Performance of Network5 for both the feature extraction methods

    Table 10: Simulated Results for NewPR model with Levenberg-Marquardt learning rule

    Description Pattern data training regressionvalueAverage of regression value test

    data samplesNetwork 6 using Feature Extraction

    method 1 0.485805 0.343696

    Network 6 using Feature Extraction

    method 20.846857 0.396131

    Figure 10: Performance of Network6 for both the feature extraction methods

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    17/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    70

    Table 11: Simulated Results for Newrbe model with Radial Basis Function Approximation

    Description Pattern data trainingregression valueAverage of regression value test

    data samplesNetwork 7 using Feature Extraction

    method 1 1 0.403733Network 7 using Feature Extraction

    method 2 1 0.112004

    Figure 11: Performance of Network7 for both the feature extraction methods

    Table 12: Simulated Results for Newrb model with Radial Basis Function Approximation

    Description Pattern data training

    regression value

    Average of regression

    value test data samplesNetwork 8 using Feature Extraction

    method 1 1 0.303037

    Network 8 using Feature Extractionmethod 2 1 0.112487

    Figure 12: Performance of Network 8 for both the feature extraction methods

    The simulation results of training are indicating that the performance of network models withRadial Basis function approximation is better than network models with Levenberg-Marquardt

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    18/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    71

    Backpropagation learning technique for the second feature extraction method i.e. each pixel value ofthe resize and processed image. Now we evaluate the performance of these trained neural networkmodels for reorganization of handwritten English capital and small alphabets, those did not presentduring the training. The performances of these networks are presented in table 13 and table 14. Thetable 13 is presenting the performance of all the six neural network models for the prototype inputpatterns processed with first method of feature extraction whereas the table 14 is presenting theperformance of all he six neural network models for the same input patterns processed with secondmethod of feature extraction. The first row of both the tables is representing the rate of correctrecognition for the presented input patterns. The second row of both the tables is presenting thecorrect number of recognized pattern among the presented arbitrary patterns.

    Table 13: Performance of all the six models for pattern recognition of presented prototype inputpatterns using first method of feature extraction

    Description newff newfit newgrnn newpr newrbe newrb

    % of charactersrecognized 10 0 10 20 25 30

    Total no. ofcharactersrecognized

    2 0 2 4 5 6

    PresentedPrototypePatterns

    Correct and Incorrect Recognition

    newff newfit newgrnn newpr netrbe newrb

    e

    j

    k

    m

    n

    p

    q

    t

    u

    v

    B

    E

    H

    J

    K

    L

    R

    X

    Y

    Z

    From the table 13 it can observer that the performance of Radial Basis function neuralnetwork is better than the other neural networks models. Its performance is even better than the exactradial basis function network. It correctly recognized 6 out of 20 prototype arbitrary input patterns ofhandwritten English alphabets. These patterns did not use in the training set and selected as thesamples of test patterns.

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    19/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    72

    Table 14: Performance of all the six models for pattern recognition of presented prototype inputpatterns using second method of feature extraction

    newff newfit newgrnn newpr newrbe newrb

    %. ofcharactersrecognized

    5 0 85 5 15 15

    Total no.of

    charactersrecognized

    1 0 17 1 3 3

    e

    j

    k

    m

    n

    p

    q

    t

    u

    v

    B

    E

    H

    J

    K

    L

    R

    X

    Y

    Z

    From the table 14 it can observer that the performance of generalized neural network modeltrained with Radial Basis function approximation is better than the other neural networks models. Itsperformance is even better than the exact radial basis function network and Radial basis Network. Itcorrectly recognized 17 out of 20 prototype arbitrary input patterns of handwritten English alphabets.It is quit noticeable that the performance of neural network is better for second method featureextraction i.e. each pixel value of the resize image only for generalized neural network with radialbasis function approximation whereas the performance of other neural network models is better forfirst method of feature extraction i.e. mean value of pixel of processed image.

    6. CONCLUSION

    This paper presented the performance evaluation of six different models of feed forwardneural networks trained with Levenberg-Marquardt Backpropagation learning technique and Radialbasis function approximation for the handwritten curve script of capital and small English alphabets.There are two feature extractions method used. In the first method the row wise mean of theproceeded image of alphabets is considered and in second method each pixel value of the resize and

    precede image is considered. The simulated results are indicating that the generalized neuralnetwork trained with radial basis function approximation for second method of feature extraction

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    20/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    73

    yields the highest rate of recognition i.e. 85% for randomly chosen 10 lower case and 10 uppercasecharacters. The remaining models of neural networks are showing poor performance irrespective ofthe feature extraction method. The following observations are considered from the simulation ofperformance evaluation:1. First method of feature extraction uses 30 features for each character whereas second method

    of feature extraction uses 100 features for the each character. Thus, it seems that more thenumber of features more is the accuracy level as far as generalized neural network model isconcern.

    2. In the training process the regression value for Radial basis network is found perfect but duringthe validation for the test pattern the performance degrades rapidly. Thus the network is welltuned for the training set but not able to generalize the behavior. It is working as goodapproximation and bad generalization.

    3. The second method of feature extraction is providing more feature values in the patterninformation with respect to the first method of feature extraction. Therefore, the performanceof each neural network model is found better for the second feature extraction method.

    7. REFERENCES

    1. J. Pradeep, E. Srinivasan and S. Himavathi, “Diagonal based feature extraction forhandwritten alphabets recognition system using neural network”, International Journal ofComputer Science & Information Technology (IJCSIT), 3 (1) 27-38 (2011)

    2. R. Plamondon and S. N. Srihari, “On-Line and Off-Line handwriting recognition recognition– A complete survey”, IEEE Transaction on pattern Recognition and Machine Intelligence,22 (1) 63-84 (2000)

    3. Kauleshwar Prasad, D. C. Nigam, Ashmika Lakhotiya and Dheeren Umre, “Character

    Recognition Using Matlab’s Neural Network Toolbox”, International Journal of u- and e-Service, Science and Technology, 6 (1) 13-20 (2013)4. Ankit Sharma and Dipti R Chaudhary, “Character Recognition Using Neural Network”,

    International Journal of Engineering Trends and Technology (IJETT), 4 (4) 662-667 (2013)5. Chirag I. Patel, Ripal Patel and Palak Patel, “Handwritten Character Recognition using

    Neural Network”, International Journal of Scientific & Engineering Research, 2 (5) 1-6(2011)

    6. Manish Mangal and Manu Pratap Singh, “Handwritten English vowels recognition usinghybrid evolutionary Feed-forward neural network”, Malaysian Journal of Computer Science,19 (2) 169-187 (2006).

    7. Anita pal and Dayashankar Singh, “Handwritten english character recognition using neural

    network”, International Journal of Computer Science & Communication, 1 (2) 141-144(2010).8. K. Y. Rajput and Sangeeta Mishra, “Recognition and editing of Devnagri handwriting using

    neural network”, Proc33dings of SPIT-IEEE Colloquium and International Conference,Mumbai, India, 1 66-70 (2008)

    9. Meenakshi Sharma and Kavita Khanna, “Offline signature verification using supervised andunsupervised neural networks”, International Journal of Computer Science and MobileComputing, 3 (7) 425-436 (2014).

    10. Priyanka Sharma and Manavjeet Kaur, “Classification in Pattern Recognition: A Review“,International Journal of Advanced Research in Computer Science and Software Engineering,3 (4) 298-306 (2013)

    11. K. Fukushima and N. Wake, "Handwritten alphanumeric character recognition by theneocognitron.", IEEE Trans. on Neural Networks, 2 (3) 355-365 (1991).

  • 8/9/2019 Handwritten Character Recognition Using Feed-Forward Neural Network Models

    21/21

    International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),ISSN 0976 - 6375(Online), Volume 6, Issue 2, February (2015), pp. 54-74 © IAEME

    74

    12. Y. L. Cun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbad and L. D. Jackel,"Handwritten digit recognition with a Backpropagation network", Neural InformationProcessing Systems, Touretzky editor, Morgan Kaufmann Publishers, (2) 396-404 (1990).

    13. A. K. Jain, J. Mao and K. M. Mohiuddin, “Artificial Neural Networks: A Tutorial.”,Computer, 31-44, (1996).

    14. B. Ripley, “Statistical Aspects of Neural Networks.”, Networks on Chaos: Statistical andProbabilistic Aspects. U. Bornndorff-Nielsen, J. Jensen, and W. Kendal, eds., Chapman andHall, (1993).

    15. J. Anderson, A. Pellionisz and E. Rosenfeld, “Neuro-computing 2: Directions for Research”,Cambridge Mass.: MIT Press, (1990).

    16. F. Rosenblatt, “Principles of Neurodynamics: Perceptron and the Theory of BrainMechanisms”, Spartan Books, Washington, D.C., (1962).

    17. B. Widrow and M. A. Lehr, “30 years of adaptive neural networks: perceptron, Madeline, andBackpropagation.”, Proceedings of the IEEE 78 (9) 1415-1442 (1990).

    18. M. L. Minsky and S. A. Papert, “Perceptron.”, Cambridge, MA: MIT Press. ExpandedEdition, (1990).

    19. S. B. Cho, “Fusion of neural networks with fuzzy logic and genet ic algorithm”, IOS Press,363–372 (2002).

    20. B. Widrow and M. E. Hoff, ”Adaptive switching circuits” IRE Eastern Electronic Show &Convention (WESCON1960), Convention Record, (4) 96–104 (1960).

    21. P. J. Werbos, “Beyond regressions: New tools for prediction and analysis in the behavioralsciences”, PhD Thesis, Harvard University, Cambridge, MA, (1974).

    22. F. J. Pineda, “Generalization of back-propagation to recurrent neural networks”, Physical RevLetter, (59) 2229–2232 (1987).

    23. R. Battiti, and F. Masulli, “BFGS optimization for faster automated supervised learning”, In:

    Proc. Int. Neural Network Conf. France, (2) 757-760 (1990)24. D. E. Rumelhart, G. E. Hinton and R. J. Williams, “Learning internal representations by errorpropagation”, MIT Press, Cambridge, (1) 318–362 (1986).

    25. P. Muneesawang and L. Guan, "Image retrieval with embedded sub-class information usingGaussian mixture models", Proceedings of International Conference on Multimedia andExpo, (2003).

    26. S. Lee. “Off-Line Recognition of Totally Unconstrained Handwritten Numerals UsingMultilayer Cluster Neural Network”, IEEE Trans. Pattern Anal. Mach. Intell. 18 (6) 648-652(1996).

    27. S. Shrivastava, S. and Manu Paratp Singh, “Performance evaluation of feed-forwardneural network with soft computing techniques for hand written English alphabets”,

    Journal of Applied Soft Computing, Elsevier, (11) 1156-1182 (2011).28. V. Subba Ramaiah and R. Rajeswara Rao, “Automatic Text-Independent Speaker TrackingSystem Using Feed-Forward Neural Networks (FFNN)” International journal of ComputerEngineering & Technology (IJCET), Volume 5, Issue 1, 2014, pp. 11 - 20, ISSN Print: 0976– 6367, ISSN Online: 0976 – 6375.

    29. M. M. Kodabagi, S. A. Angadi and Chetana. R. Shivanagi, “Character Recognition ofKannada Text In Scene Images Using Neural Network” International Journal Of GraphicsAnd Multimedia (IJGM), Volume 4, Issue 1, 2014, pp. 9 - 19, ISSN Print: 0976 – 6448, ISSNOnline: 0976 –6456.

    30. Ms. Aruna J. Chamatkar and Dr. P.K. Butey, “Performance Analysis of Data MiningAlgorithms with Neural Network” International journal of Computer Engineering &

    Technology (IJCET), Volume 6, Issue 1, 2015, pp. 1 - 11, ISSN Print: 0976 – 6367, ISSNOnline: 0976 – 6375.