SlideShare uma empresa Scribd logo
1 de 3
Baixar para ler offline
Classification of Handwritten Characters by their
Symmetry Features
Sam Holland and Richard Neville
Machine Learning & Optimisation Research Group
School of Computer Science
University of Manchester
Email: sam.holland@postgrad.manchester.ac.uk, r.neville@manchester.ac.uk
Abstract—We propose a technique to classify characters by
two different forms of their symmetry features. The Generalized
Symmetry Transform is applied to digits from the USPS data
set. These features are then used to train Probabilistic Neural
Networks and their performances are compared to the traditional
method.
I. INTRODUCTION
Symmetry is a property of many of the objects in the
world around us, often a by-product of simplicity in design
and manufacturing. Symmetry is not widely used to assist
Artificial Neural Networks (ANNs) to perform classification.
The performance of Optical Character Recognition (OCR)
systems has not yet reached that of humans. We believe that
symmetry features used by Neural Networks will improve
recogntion.
The aim of this work is to test the hypotheses that symmetry
features of patterns can be used for classification; and that
classification of a pattern’s symmetry features can improve
generalization. To this end, we trained with a large set of
handwritten characters collected from real world samples. We
partitioned the set into two groups: a training set and a test set.
The performance on the test set provides an insight into the
net’s ability to generalize - i.e. correctly classify new, unseen
patterns which vary from those in the training set.
II. PROBABILISTIC NEURAL NETWORKS
Probabilistic Neural Networks (PNNs) have been shown to
outperform the traditional Multi-Layer Perceptron (MLP) [2],
[3] classifier in Optical Character Recognition (OCR) tasks [7].
Their main advantage is that their training is trivial and near
instantaneous [1], [5], requiring only a single pass. Training is
based on estimating probability distribution functions utilising
Parzen windows. They can support very complex decision
surfaces which approach the Bayes optimal [6].
Probabilistic Neural Networks have three layers of neurons
(not including an input layer): the pattern layer; the summation
layer; and the output layer. They are connected in a feed-
forward manner. The structure is depicted in Figure 1.
A. Pattern Units
The pattern units are similar to the semi-linear neurons of
an MLP except that they use an exponential activation-output
function:
g(Zi) = e
Zi−1
σ2
(1)
Fig. 1. The structure of a probabilistic neural network.
A smoothing parameter, σ, identical for all the units in the
layer, controls the exponential scale factor. The activation, Zi
is defined as the sum of the weighted (W) inputs (X):
Zi = X · Wi (2)
For each training pattern there is one pattern unit. The
training pattern input is used as the weight for that pattern unit.
Whilst this layer will be large for any non-trivial classification
task, training is completed within a single pass. Recognition
is slower, but training will be quicker than an MLP used for
the same task.
After training, the output of these units represents how
similar the new pattern is to the training patterns. The unit
with the highest output represents the closest matching training
pattern.
B. Summation Units
The summation units simply sum the output of pattern units
belonging to the same class:
fA(X) =
i∈A
Zi (3)
C. Output Units
The output units have two inputs and a binary output. The
first input is from the summation unit of patterns belonging to
that class. The second input is from a summation unit of non-
class members. This input is inverted (multiplied by negative
2009 International Conference on Advances in Computing, Control, and Telecommunication Technologies
978-0-7695-3915-7/09 $26.00 © 2009 IEEE
DOI 10.1109/ACT.2009.85
316
one). Hence, the activation for the output unit is equivalent to:
a = fA(X) − fB(X) (4)
The activation-output function is a hard limiter:
y =
+1 if a ≥ 0
−1 if a < 0
(5)
D. PNN Smoothing Parameter, σ
A PNN has only one parameter (other than the training
patterns): The smoothing parameter, σ, affects the network’s
Probability Distribution Function (PDF) estimate. As σ → 0,
the network acts as a nearest-neighbour classifier. As σ → ∞,
the network creates a non-linear decision boundary approach-
ing the Bayes optimal. In practice, it is not difficult to find
the optimal value for σ and the value can be modified after
training of the weights is complete.
III. THE GENERALIZED SYMMETRY TRANSFORM
Reisfeld et al [4] proposed a method of measuring symmetry
called the Generalized Symmetry Transform. The measure
compares pairs of edge point intensities and orientations. The
GST produces a detailed map of the symmetry magnitudes (as
a continuous measure) and symmetry orientations, rather than
produce an average symmetry position.
Identifying the location of just one reflection plane, that acts
globally on an object, is of limited use to classification. Two or
more classes of characters are likely to share such a reflection
plane. A classifier would be unable to distinguish between
these classes. The GST was selected for this experiment
because it identifies symmetries which are present in only a
subsection of the object. We call this reflectional relationship
between a subset of pixels local symmetry.
The symmetry magnitude is calculated from an approxima-
tion of the pixel greyscale intensity gradients. We can approx-
imate the gradient using the Sobel edge operator convolution
masks:
∆pk =
δ
δx
pk,
δ
δy
pk (6)
A vector, vk = (rk, θk) is associated with each pixel k
in the input image consisting of a magnitude of the intensity
gradient, rk and the angle between the normal of the gradient
and the horizontal axis, θk.
rk = log(1 + ∆pk ) (7)
θk = arctan
δ
δy
pk
δ
δx
pk
(8)
Each pair of pixels contributes to the symmetry magnitude
at a pixel mid-way between them. Hence, the set Γ(p) consists
of all the pixel-pairs that contribute to the magnitude at a point
p:
Γ(p) = (i, j)|
pi + pj
2
= p (9)
Fig. 2. Point pairs contribute to a symmetry magnitude and direction at a
point mid-way between them.
Fig. 3. The phase function of the Generalized Symmetry Transform, for
input edge orientations from two points, in the range 0 to 2π radians.
A Gaussian-based distance weight function favours contri-
butions from pixel-pairs that are closer together than those
further apart:
Dµ(i, j) =
1
√
2πµ
e
−
pi+pj
2µ
(10)
The shape of the gaussian is controlled by µ. A phase weight
function favours opposing gradient orientations (rather than
aligned orientations) from a pair of pixels:
P(i, j) = [1 − cos(θi + θj − 2αij)] × [1 − cos(θi − θj)]
(11)
Where αij is the angle between a line connection points i
and j, and the horizontal axis. Figure 3 shows the effect of
the two edge orientations on the symmetry measure. Whilst
the other components of the symmetry contribution calculate
a significance (equivalent to weighting), the phase function
is the continuous measure of symmetry. Two orientations
are symmetrical if θi − θj = ±π. Two orientations are not
symmetrical if θi − θj = 0.
The total contribution of two pixel points is a function of
317
the distance and phase weight functions:
C(i, j) = Dµ(i, j)P(i, j)rirj (12)
The symmetry magnitude at a given point p is the sum of
the contributions over a set of pixels pairs, Γ(p):
Mµ(p) =
(i,j)∈Γ(p)
C(i, j) (13)
The contribution has a direction, which is calculated from
the two contribution gradient orientations:
ϕ(i, j) =
θi + θj
2
(14)
The direction of the symmetry at a specific point p is φ(p) =
ϕ(i, j) such that C(i, j) is the maximum in the set Γ(p).
IV. EXPERIMENT
The experiment used the United States Postal Service
(USPS) data set. The set contains 9292 patterns each represent-
ing a handwritten digit (the numerals zero to nine inclusive).
A 16-by-16 grayscale image describes each digit. We trained
with networks with 7291 patterns. The networks’ performance
was measured by presenting 2001 previously unseen patterns.
One network trained with the original patterns without
any additional processing. A second network trained on the
network symmetry magnitude map, Mµ and a third trained on
the symmetry orientation map, φ.
We normalised the magnitude maps into the range −1 to 1.
The GST represents no symmetry with a zero value, but these
do not propagate well through weighted-sum connectionist
systems.
We varied the Probabilistic Neural Networks’ smoothing
parameter and measured the effect on the classification per-
formance to find the optimal performance.
V. RESULTS
All three networks classify 100% of the training patterns.
The optimal performances were: 95.17% at σ = 0.5 using
the original data as input; 87.2% at σ = 0.1 for the symmetry
magnitude map; and 72.2% at σ = 0.7 for the symmetry orien-
tation map. The relationship between the smoothing operator
and the recognition rate for all three networks is shown in
Figure 4 and listed in Table I.
VI. DISCUSSION
Both symmetry feature based networks classified 100%
of the training data. Hence, handwritten characters can be
classified by their symmetry features. However, the network
using the unprocessed USPS data set outperforms the networks
using the symmetry magnitude and orientation maps in terms
of generalization. The smoothing operator value indicates
that all sets of data are better suited to nearest-neighbour
classification, suggesting no clear separation of classes within
the decision space.
Only one form of symmetry was used in this experiment:
reflectional (or mirror) symmetry. It it too early to discount
symmetry’s potential significance. Research into the use of
Fig. 4. Graph showing the relationship between the PNN smoothing
parameter and recognition rate of the three networks when generalizing.
TABLE I
THE RELATIONSHIP BETWEEN THE PNN SMOOTHING PARAMETER, σ,
AND THE RECOGNITION RATE FOR EACH OF THE THREE INPUT TYPES.
σ
Recognition Rate
Original Magnitude Orientation
0.1 53.26 87.24 28.70
0.2 92.72 86.85 56.10
0.3 95.02 82.06 79.02
0.4 95.12 73.24 79.17
0.5 95.17 64.52 79.17
0.6 95.17 57.65 79.12
0.7 95.17 50.92 79.22
0.8 95.02 45.89 79.22
0.9 94.97 41.90 79.02
rotational (or combinations of) symmetry should be conducted.
Normalisation techniques applied to the magnitude map or ori-
entation map prior to classification may improve performance.
ACKNOWLEDGMENT
This research is funded by the Engineering and Physical
Sciences Research Council (EPSRC).
REFERENCES
[1] F. Ancona, A.M. Colla, S. Rovetta, and R. Zunino. Implementing
probabilistic neural networks. In Neural Computing and Applications,
volume 5, pages 152–159, 1997.
[2] G.E. Hinton, D.E. Rumelhart, and R.J. Williams. Learning internal
representations by error propagation, volume 1. 1986.
[3] Y. LeCun, I. Guyon, L.D. Jackel, D. Henderson, B. Boser, R.E. Howard,
J.S. Denker, W. Hubbard, and H.P. Graf. Handwritten digit recognition:
applications of neural network chips and automatic learning. IEEE
Communications Magazine, 11:41–61, 1989.
[4] Daniel Reisfeld, Haim Wolfson, and Yehezkel Yeshurun. Context free
attentional operators: the generalized symmetry transform. International
Journal of Computer Vision, 14:119–130, 1995.
[5] D.F. Specht. Probabilistic neural networks for classification, mapping
or associative memories. In IEEE International Conference on Neural
Networks, volume 1, pages 525–532, 1988.
[6] D.F. Specht. Probabilistic neural networks. In Neural Networks, volume 3,
pages 109–118, 1990.
[7] D.F. Specht. Probabilistic neural networks and general regression neural
networks, pages 3.1–3.44. McGraw-Hill Inc, New York, 1996.
318

Mais conteúdo relacionado

Mais procurados

(研究会輪読) Facial Landmark Detection by Deep Multi-task Learning
(研究会輪読) Facial Landmark Detection by Deep Multi-task Learning(研究会輪読) Facial Landmark Detection by Deep Multi-task Learning
(研究会輪読) Facial Landmark Detection by Deep Multi-task Learning
Masahiro Suzuki
 
Image recogonization
Image recogonizationImage recogonization
Image recogonization
SANTOSH RATH
 
Lecture 18: Gaussian Mixture Models and Expectation Maximization
Lecture 18: Gaussian Mixture Models and Expectation MaximizationLecture 18: Gaussian Mixture Models and Expectation Maximization
Lecture 18: Gaussian Mixture Models and Expectation Maximization
butest
 

Mais procurados (20)

Tracking Faces using Active Appearance Models
Tracking Faces using Active Appearance ModelsTracking Faces using Active Appearance Models
Tracking Faces using Active Appearance Models
 
The step construction of penalized spline in electrical power load data
The step construction of penalized spline in electrical power load dataThe step construction of penalized spline in electrical power load data
The step construction of penalized spline in electrical power load data
 
Principal component analysis
Principal component analysisPrincipal component analysis
Principal component analysis
 
Spectral cnn
Spectral cnnSpectral cnn
Spectral cnn
 
A Correlative Information-Theoretic Measure for Image Similarity
A Correlative Information-Theoretic Measure for Image SimilarityA Correlative Information-Theoretic Measure for Image Similarity
A Correlative Information-Theoretic Measure for Image Similarity
 
(研究会輪読) Facial Landmark Detection by Deep Multi-task Learning
(研究会輪読) Facial Landmark Detection by Deep Multi-task Learning(研究会輪読) Facial Landmark Detection by Deep Multi-task Learning
(研究会輪読) Facial Landmark Detection by Deep Multi-task Learning
 
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATION
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATIONA STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATION
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATION
 
Lda
LdaLda
Lda
 
Application of normalized cross correlation to image registration
Application of normalized cross correlation to image registrationApplication of normalized cross correlation to image registration
Application of normalized cross correlation to image registration
 
An efficient approach to wavelet image Denoising
An efficient approach to wavelet image DenoisingAn efficient approach to wavelet image Denoising
An efficient approach to wavelet image Denoising
 
Subspace based doa estimation techniques
Subspace based doa estimation techniquesSubspace based doa estimation techniques
Subspace based doa estimation techniques
 
Direction of arrival estimation using music algorithm
Direction of arrival estimation using music algorithmDirection of arrival estimation using music algorithm
Direction of arrival estimation using music algorithm
 
Zoooooohaib
ZoooooohaibZoooooohaib
Zoooooohaib
 
The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)
 
Image recogonization
Image recogonizationImage recogonization
Image recogonization
 
Markov Cluster Algorithm & real world application
Markov Cluster Algorithm & real world applicationMarkov Cluster Algorithm & real world application
Markov Cluster Algorithm & real world application
 
COMPARISON OF VOLUME AND DISTANCE CONSTRAINT ON HYPERSPECTRAL UNMIXING
COMPARISON OF VOLUME AND DISTANCE CONSTRAINT ON HYPERSPECTRAL UNMIXINGCOMPARISON OF VOLUME AND DISTANCE CONSTRAINT ON HYPERSPECTRAL UNMIXING
COMPARISON OF VOLUME AND DISTANCE CONSTRAINT ON HYPERSPECTRAL UNMIXING
 
IMAGE REGISTRATION USING ADVANCED TOPOLOGY PRESERVING RELAXATION LABELING
IMAGE REGISTRATION USING ADVANCED TOPOLOGY PRESERVING RELAXATION LABELING IMAGE REGISTRATION USING ADVANCED TOPOLOGY PRESERVING RELAXATION LABELING
IMAGE REGISTRATION USING ADVANCED TOPOLOGY PRESERVING RELAXATION LABELING
 
Lecture 18: Gaussian Mixture Models and Expectation Maximization
Lecture 18: Gaussian Mixture Models and Expectation MaximizationLecture 18: Gaussian Mixture Models and Expectation Maximization
Lecture 18: Gaussian Mixture Models and Expectation Maximization
 
Real interpolation method for transfer function approximation of distributed ...
Real interpolation method for transfer function approximation of distributed ...Real interpolation method for transfer function approximation of distributed ...
Real interpolation method for transfer function approximation of distributed ...
 

Semelhante a Classification of handwritten characters by their symmetry features

A simple framework for contrastive learning of visual representations
A simple framework for contrastive learning of visual representationsA simple framework for contrastive learning of visual representations
A simple framework for contrastive learning of visual representations
Devansh16
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
IJERD Editor
 
A comparison-of-first-and-second-order-training-algorithms-for-artificial-neu...
A comparison-of-first-and-second-order-training-algorithms-for-artificial-neu...A comparison-of-first-and-second-order-training-algorithms-for-artificial-neu...
A comparison-of-first-and-second-order-training-algorithms-for-artificial-neu...
Cemal Ardil
 
Face recognition using laplacianfaces (synopsis)
Face recognition using laplacianfaces (synopsis)Face recognition using laplacianfaces (synopsis)
Face recognition using laplacianfaces (synopsis)
Mumbai Academisc
 
Drobics, m. 2001: datamining using synergiesbetween self-organising maps and...
Drobics, m. 2001:  datamining using synergiesbetween self-organising maps and...Drobics, m. 2001:  datamining using synergiesbetween self-organising maps and...
Drobics, m. 2001: datamining using synergiesbetween self-organising maps and...
ArchiLab 7
 

Semelhante a Classification of handwritten characters by their symmetry features (20)

Joint3DShapeMatching
Joint3DShapeMatchingJoint3DShapeMatching
Joint3DShapeMatching
 
Joint3DShapeMatching - a fast approach to 3D model matching using MatchALS 3...
Joint3DShapeMatching  - a fast approach to 3D model matching using MatchALS 3...Joint3DShapeMatching  - a fast approach to 3D model matching using MatchALS 3...
Joint3DShapeMatching - a fast approach to 3D model matching using MatchALS 3...
 
Investigation on the Pattern Synthesis of Subarray Weights for Low EMI Applic...
Investigation on the Pattern Synthesis of Subarray Weights for Low EMI Applic...Investigation on the Pattern Synthesis of Subarray Weights for Low EMI Applic...
Investigation on the Pattern Synthesis of Subarray Weights for Low EMI Applic...
 
Dycops2019
Dycops2019 Dycops2019
Dycops2019
 
A simple framework for contrastive learning of visual representations
A simple framework for contrastive learning of visual representationsA simple framework for contrastive learning of visual representations
A simple framework for contrastive learning of visual representations
 
50120130406039
5012013040603950120130406039
50120130406039
 
IMAGE REGISTRATION USING ADVANCED TOPOLOGY PRESERVING RELAXATION LABELING
IMAGE REGISTRATION USING ADVANCED TOPOLOGY PRESERVING RELAXATION LABELING IMAGE REGISTRATION USING ADVANCED TOPOLOGY PRESERVING RELAXATION LABELING
IMAGE REGISTRATION USING ADVANCED TOPOLOGY PRESERVING RELAXATION LABELING
 
Comparison of image segmentation
Comparison of image segmentationComparison of image segmentation
Comparison of image segmentation
 
Feed forward neural network for sine
Feed forward neural network for sineFeed forward neural network for sine
Feed forward neural network for sine
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
A comparison-of-first-and-second-order-training-algorithms-for-artificial-neu...
A comparison-of-first-and-second-order-training-algorithms-for-artificial-neu...A comparison-of-first-and-second-order-training-algorithms-for-artificial-neu...
A comparison-of-first-and-second-order-training-algorithms-for-artificial-neu...
 
An ann approach for network
An ann approach for networkAn ann approach for network
An ann approach for network
 
Blind Image Seperation Using Forward Difference Method (FDM)
Blind Image Seperation Using Forward Difference Method (FDM)Blind Image Seperation Using Forward Difference Method (FDM)
Blind Image Seperation Using Forward Difference Method (FDM)
 
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
 
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
 
Face recognition using laplacianfaces (synopsis)
Face recognition using laplacianfaces (synopsis)Face recognition using laplacianfaces (synopsis)
Face recognition using laplacianfaces (synopsis)
 
Exact network reconstruction from consensus signals and one eigen value
Exact network reconstruction from consensus signals and one eigen valueExact network reconstruction from consensus signals and one eigen value
Exact network reconstruction from consensus signals and one eigen value
 
2009 asilomar
2009 asilomar2009 asilomar
2009 asilomar
 
9.pdf
9.pdf9.pdf
9.pdf
 
Drobics, m. 2001: datamining using synergiesbetween self-organising maps and...
Drobics, m. 2001:  datamining using synergiesbetween self-organising maps and...Drobics, m. 2001:  datamining using synergiesbetween self-organising maps and...
Drobics, m. 2001: datamining using synergiesbetween self-organising maps and...
 

Último

Hospital management system project report.pdf
Hospital management system project report.pdfHospital management system project report.pdf
Hospital management system project report.pdf
Kamal Acharya
 
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
AldoGarca30
 
DeepFakes presentation : brief idea of DeepFakes
DeepFakes presentation : brief idea of DeepFakesDeepFakes presentation : brief idea of DeepFakes
DeepFakes presentation : brief idea of DeepFakes
MayuraD1
 

Último (20)

DC MACHINE-Motoring and generation, Armature circuit equation
DC MACHINE-Motoring and generation, Armature circuit equationDC MACHINE-Motoring and generation, Armature circuit equation
DC MACHINE-Motoring and generation, Armature circuit equation
 
Hospital management system project report.pdf
Hospital management system project report.pdfHospital management system project report.pdf
Hospital management system project report.pdf
 
Unleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapUnleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leap
 
School management system project Report.pdf
School management system project Report.pdfSchool management system project Report.pdf
School management system project Report.pdf
 
Bhubaneswar🌹Call Girls Bhubaneswar ❤Komal 9777949614 💟 Full Trusted CALL GIRL...
Bhubaneswar🌹Call Girls Bhubaneswar ❤Komal 9777949614 💟 Full Trusted CALL GIRL...Bhubaneswar🌹Call Girls Bhubaneswar ❤Komal 9777949614 💟 Full Trusted CALL GIRL...
Bhubaneswar🌹Call Girls Bhubaneswar ❤Komal 9777949614 💟 Full Trusted CALL GIRL...
 
Introduction to Serverless with AWS Lambda
Introduction to Serverless with AWS LambdaIntroduction to Serverless with AWS Lambda
Introduction to Serverless with AWS Lambda
 
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
 
Online electricity billing project report..pdf
Online electricity billing project report..pdfOnline electricity billing project report..pdf
Online electricity billing project report..pdf
 
data_management_and _data_science_cheat_sheet.pdf
data_management_and _data_science_cheat_sheet.pdfdata_management_and _data_science_cheat_sheet.pdf
data_management_and _data_science_cheat_sheet.pdf
 
Design For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startDesign For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the start
 
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...
 
DeepFakes presentation : brief idea of DeepFakes
DeepFakes presentation : brief idea of DeepFakesDeepFakes presentation : brief idea of DeepFakes
DeepFakes presentation : brief idea of DeepFakes
 
457503602-5-Gas-Well-Testing-and-Analysis-pptx.pptx
457503602-5-Gas-Well-Testing-and-Analysis-pptx.pptx457503602-5-Gas-Well-Testing-and-Analysis-pptx.pptx
457503602-5-Gas-Well-Testing-and-Analysis-pptx.pptx
 
Work-Permit-Receiver-in-Saudi-Aramco.pptx
Work-Permit-Receiver-in-Saudi-Aramco.pptxWork-Permit-Receiver-in-Saudi-Aramco.pptx
Work-Permit-Receiver-in-Saudi-Aramco.pptx
 
FEA Based Level 3 Assessment of Deformed Tanks with Fluid Induced Loads
FEA Based Level 3 Assessment of Deformed Tanks with Fluid Induced LoadsFEA Based Level 3 Assessment of Deformed Tanks with Fluid Induced Loads
FEA Based Level 3 Assessment of Deformed Tanks with Fluid Induced Loads
 
Thermal Engineering -unit - III & IV.ppt
Thermal Engineering -unit - III & IV.pptThermal Engineering -unit - III & IV.ppt
Thermal Engineering -unit - III & IV.ppt
 
NO1 Top No1 Amil Baba In Azad Kashmir, Kashmir Black Magic Specialist Expert ...
NO1 Top No1 Amil Baba In Azad Kashmir, Kashmir Black Magic Specialist Expert ...NO1 Top No1 Amil Baba In Azad Kashmir, Kashmir Black Magic Specialist Expert ...
NO1 Top No1 Amil Baba In Azad Kashmir, Kashmir Black Magic Specialist Expert ...
 
💚Trustworthy Call Girls Pune Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top...
💚Trustworthy Call Girls Pune Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top...💚Trustworthy Call Girls Pune Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top...
💚Trustworthy Call Girls Pune Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top...
 
Double Revolving field theory-how the rotor develops torque
Double Revolving field theory-how the rotor develops torqueDouble Revolving field theory-how the rotor develops torque
Double Revolving field theory-how the rotor develops torque
 
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...
 

Classification of handwritten characters by their symmetry features

  • 1. Classification of Handwritten Characters by their Symmetry Features Sam Holland and Richard Neville Machine Learning & Optimisation Research Group School of Computer Science University of Manchester Email: sam.holland@postgrad.manchester.ac.uk, r.neville@manchester.ac.uk Abstract—We propose a technique to classify characters by two different forms of their symmetry features. The Generalized Symmetry Transform is applied to digits from the USPS data set. These features are then used to train Probabilistic Neural Networks and their performances are compared to the traditional method. I. INTRODUCTION Symmetry is a property of many of the objects in the world around us, often a by-product of simplicity in design and manufacturing. Symmetry is not widely used to assist Artificial Neural Networks (ANNs) to perform classification. The performance of Optical Character Recognition (OCR) systems has not yet reached that of humans. We believe that symmetry features used by Neural Networks will improve recogntion. The aim of this work is to test the hypotheses that symmetry features of patterns can be used for classification; and that classification of a pattern’s symmetry features can improve generalization. To this end, we trained with a large set of handwritten characters collected from real world samples. We partitioned the set into two groups: a training set and a test set. The performance on the test set provides an insight into the net’s ability to generalize - i.e. correctly classify new, unseen patterns which vary from those in the training set. II. PROBABILISTIC NEURAL NETWORKS Probabilistic Neural Networks (PNNs) have been shown to outperform the traditional Multi-Layer Perceptron (MLP) [2], [3] classifier in Optical Character Recognition (OCR) tasks [7]. Their main advantage is that their training is trivial and near instantaneous [1], [5], requiring only a single pass. Training is based on estimating probability distribution functions utilising Parzen windows. They can support very complex decision surfaces which approach the Bayes optimal [6]. Probabilistic Neural Networks have three layers of neurons (not including an input layer): the pattern layer; the summation layer; and the output layer. They are connected in a feed- forward manner. The structure is depicted in Figure 1. A. Pattern Units The pattern units are similar to the semi-linear neurons of an MLP except that they use an exponential activation-output function: g(Zi) = e Zi−1 σ2 (1) Fig. 1. The structure of a probabilistic neural network. A smoothing parameter, σ, identical for all the units in the layer, controls the exponential scale factor. The activation, Zi is defined as the sum of the weighted (W) inputs (X): Zi = X · Wi (2) For each training pattern there is one pattern unit. The training pattern input is used as the weight for that pattern unit. Whilst this layer will be large for any non-trivial classification task, training is completed within a single pass. Recognition is slower, but training will be quicker than an MLP used for the same task. After training, the output of these units represents how similar the new pattern is to the training patterns. The unit with the highest output represents the closest matching training pattern. B. Summation Units The summation units simply sum the output of pattern units belonging to the same class: fA(X) = i∈A Zi (3) C. Output Units The output units have two inputs and a binary output. The first input is from the summation unit of patterns belonging to that class. The second input is from a summation unit of non- class members. This input is inverted (multiplied by negative 2009 International Conference on Advances in Computing, Control, and Telecommunication Technologies 978-0-7695-3915-7/09 $26.00 © 2009 IEEE DOI 10.1109/ACT.2009.85 316
  • 2. one). Hence, the activation for the output unit is equivalent to: a = fA(X) − fB(X) (4) The activation-output function is a hard limiter: y = +1 if a ≥ 0 −1 if a < 0 (5) D. PNN Smoothing Parameter, σ A PNN has only one parameter (other than the training patterns): The smoothing parameter, σ, affects the network’s Probability Distribution Function (PDF) estimate. As σ → 0, the network acts as a nearest-neighbour classifier. As σ → ∞, the network creates a non-linear decision boundary approach- ing the Bayes optimal. In practice, it is not difficult to find the optimal value for σ and the value can be modified after training of the weights is complete. III. THE GENERALIZED SYMMETRY TRANSFORM Reisfeld et al [4] proposed a method of measuring symmetry called the Generalized Symmetry Transform. The measure compares pairs of edge point intensities and orientations. The GST produces a detailed map of the symmetry magnitudes (as a continuous measure) and symmetry orientations, rather than produce an average symmetry position. Identifying the location of just one reflection plane, that acts globally on an object, is of limited use to classification. Two or more classes of characters are likely to share such a reflection plane. A classifier would be unable to distinguish between these classes. The GST was selected for this experiment because it identifies symmetries which are present in only a subsection of the object. We call this reflectional relationship between a subset of pixels local symmetry. The symmetry magnitude is calculated from an approxima- tion of the pixel greyscale intensity gradients. We can approx- imate the gradient using the Sobel edge operator convolution masks: ∆pk = δ δx pk, δ δy pk (6) A vector, vk = (rk, θk) is associated with each pixel k in the input image consisting of a magnitude of the intensity gradient, rk and the angle between the normal of the gradient and the horizontal axis, θk. rk = log(1 + ∆pk ) (7) θk = arctan δ δy pk δ δx pk (8) Each pair of pixels contributes to the symmetry magnitude at a pixel mid-way between them. Hence, the set Γ(p) consists of all the pixel-pairs that contribute to the magnitude at a point p: Γ(p) = (i, j)| pi + pj 2 = p (9) Fig. 2. Point pairs contribute to a symmetry magnitude and direction at a point mid-way between them. Fig. 3. The phase function of the Generalized Symmetry Transform, for input edge orientations from two points, in the range 0 to 2π radians. A Gaussian-based distance weight function favours contri- butions from pixel-pairs that are closer together than those further apart: Dµ(i, j) = 1 √ 2πµ e − pi+pj 2µ (10) The shape of the gaussian is controlled by µ. A phase weight function favours opposing gradient orientations (rather than aligned orientations) from a pair of pixels: P(i, j) = [1 − cos(θi + θj − 2αij)] × [1 − cos(θi − θj)] (11) Where αij is the angle between a line connection points i and j, and the horizontal axis. Figure 3 shows the effect of the two edge orientations on the symmetry measure. Whilst the other components of the symmetry contribution calculate a significance (equivalent to weighting), the phase function is the continuous measure of symmetry. Two orientations are symmetrical if θi − θj = ±π. Two orientations are not symmetrical if θi − θj = 0. The total contribution of two pixel points is a function of 317
  • 3. the distance and phase weight functions: C(i, j) = Dµ(i, j)P(i, j)rirj (12) The symmetry magnitude at a given point p is the sum of the contributions over a set of pixels pairs, Γ(p): Mµ(p) = (i,j)∈Γ(p) C(i, j) (13) The contribution has a direction, which is calculated from the two contribution gradient orientations: ϕ(i, j) = θi + θj 2 (14) The direction of the symmetry at a specific point p is φ(p) = ϕ(i, j) such that C(i, j) is the maximum in the set Γ(p). IV. EXPERIMENT The experiment used the United States Postal Service (USPS) data set. The set contains 9292 patterns each represent- ing a handwritten digit (the numerals zero to nine inclusive). A 16-by-16 grayscale image describes each digit. We trained with networks with 7291 patterns. The networks’ performance was measured by presenting 2001 previously unseen patterns. One network trained with the original patterns without any additional processing. A second network trained on the network symmetry magnitude map, Mµ and a third trained on the symmetry orientation map, φ. We normalised the magnitude maps into the range −1 to 1. The GST represents no symmetry with a zero value, but these do not propagate well through weighted-sum connectionist systems. We varied the Probabilistic Neural Networks’ smoothing parameter and measured the effect on the classification per- formance to find the optimal performance. V. RESULTS All three networks classify 100% of the training patterns. The optimal performances were: 95.17% at σ = 0.5 using the original data as input; 87.2% at σ = 0.1 for the symmetry magnitude map; and 72.2% at σ = 0.7 for the symmetry orien- tation map. The relationship between the smoothing operator and the recognition rate for all three networks is shown in Figure 4 and listed in Table I. VI. DISCUSSION Both symmetry feature based networks classified 100% of the training data. Hence, handwritten characters can be classified by their symmetry features. However, the network using the unprocessed USPS data set outperforms the networks using the symmetry magnitude and orientation maps in terms of generalization. The smoothing operator value indicates that all sets of data are better suited to nearest-neighbour classification, suggesting no clear separation of classes within the decision space. Only one form of symmetry was used in this experiment: reflectional (or mirror) symmetry. It it too early to discount symmetry’s potential significance. Research into the use of Fig. 4. Graph showing the relationship between the PNN smoothing parameter and recognition rate of the three networks when generalizing. TABLE I THE RELATIONSHIP BETWEEN THE PNN SMOOTHING PARAMETER, σ, AND THE RECOGNITION RATE FOR EACH OF THE THREE INPUT TYPES. σ Recognition Rate Original Magnitude Orientation 0.1 53.26 87.24 28.70 0.2 92.72 86.85 56.10 0.3 95.02 82.06 79.02 0.4 95.12 73.24 79.17 0.5 95.17 64.52 79.17 0.6 95.17 57.65 79.12 0.7 95.17 50.92 79.22 0.8 95.02 45.89 79.22 0.9 94.97 41.90 79.02 rotational (or combinations of) symmetry should be conducted. Normalisation techniques applied to the magnitude map or ori- entation map prior to classification may improve performance. ACKNOWLEDGMENT This research is funded by the Engineering and Physical Sciences Research Council (EPSRC). REFERENCES [1] F. Ancona, A.M. Colla, S. Rovetta, and R. Zunino. Implementing probabilistic neural networks. In Neural Computing and Applications, volume 5, pages 152–159, 1997. [2] G.E. Hinton, D.E. Rumelhart, and R.J. Williams. Learning internal representations by error propagation, volume 1. 1986. [3] Y. LeCun, I. Guyon, L.D. Jackel, D. Henderson, B. Boser, R.E. Howard, J.S. Denker, W. Hubbard, and H.P. Graf. Handwritten digit recognition: applications of neural network chips and automatic learning. IEEE Communications Magazine, 11:41–61, 1989. [4] Daniel Reisfeld, Haim Wolfson, and Yehezkel Yeshurun. Context free attentional operators: the generalized symmetry transform. International Journal of Computer Vision, 14:119–130, 1995. [5] D.F. Specht. Probabilistic neural networks for classification, mapping or associative memories. In IEEE International Conference on Neural Networks, volume 1, pages 525–532, 1988. [6] D.F. Specht. Probabilistic neural networks. In Neural Networks, volume 3, pages 109–118, 1990. [7] D.F. Specht. Probabilistic neural networks and general regression neural networks, pages 3.1–3.44. McGraw-Hill Inc, New York, 1996. 318