SlideShare uma empresa Scribd logo
1 de 14
Baixar para ler offline
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING &
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

TECHNOLOGY (IJCET)

ISSN 0976 – 6367(Print)
ISSN 0976 – 6375(Online)
Volume 4, Issue 6, November - December (2013), pp. 48-61
© IAEME: www.iaeme.com/ijcet.asp
Journal Impact Factor (2013): 6.1302 (Calculated by GISI)
www.jifactor.com

IJCET
©IAEME

OPTICAL CHARACTER RECOGNITION (OCR) SYSTEM FOR MULTIFONT ENGLISH TEXTS USING DCT & WAVELET TRANSFORM
Dr. MUSTAFA DHIAA AL-HASSANI
Computer Science/Mustansiriyah University, Baghdad, Iraq

ABSTRACT
Optical Character Recognition (OCR) is a type of computer software designed to translate
images of handwritten or typewritten text (usually captured by a scanner or a camera) into machineeditable text by recognizing characters at high speeds one at a time. OCR began as a field of research
in pattern recognition, artificial intelligence and machine vision. It is becoming more and more
important in the modern world according to economic reasons and business requirements. It helps
humans ease their jobs and solve more complex problems by eliminating the time-consuming spent
by human operators to re-type the documents and reduce error-prone processes.
The presence of any type of noise or a combination of them can severely degrade the
performance of OCR system. Though, a number of preprocessing techniques are considered in the
present work in order to improve the obtained accuracy of the recognized text. An OCR system for
3185 training samples and 13650 testing samples is presented for multi-font English texts.
Experiments have shown that wavelet features produce better recognition rates 96% than DCT
features 92%. An improvement overall recognition rates (about 3%) are obtained after classifying
characters according to the proportion of Height to Width feature to produce 99% for wavelet and
95% for DCT.
Keywords: DCT, Feature Extraction, Optical Character Recognition (OCR), Pattern Recognition,
Segmentation, Wavelet Transform.
1. INTRODUCTION
Since the start of the computing era, information has been represented digitally so that it can
be processed by computers. Approximately, more than 200 million paper books are being published
yearly. Paper books and documents were abundant and widely being published at that time; and
hence, there was a need to convert them into digital format. OCR was invented to translate the
traditional paper-based books into digital e-books(i.e., electronic files). It was estimated that over 2
48
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

million e-books are available for download on the Internet. E-books require less storage space than
paper books, they can also be replicated many times, shared online, and digitally processed easily,
mainly searched, translated, edited, and annotated. OCR systems are not that perfect as they are
erroneous and exhibit spelling errors in the recognized output text, especially when the images being
scanned are of poor printing quality [1, 2].
OCR is one of the most fascinating and challenging areas of pattern recognition with various
practical applications: Automated postal address reading, ZIP code reading, checks, payment slips,
income tax forms, business forms, automatic car plate-number recognition, and it can be used as an
aid for visually handicapped people when combined with speech synthesizer [1, 3, 4].
Automatic character recognition is a subfield of pattern recognition and can either to be online or off-line. On-line recognition refers to those systems where the data to be recognized is input
through a tablet digitizer, which acquires in real-time the position of the pen tip as the user writes. In
contrast, off-line systems input the data from a document through an acquisition device, such as a
scanner or a camera. Off-line character recognition is moreover divided into two categories:
machine printed and handwritten [5].
The printed texts include all the printed materials such as: books, newspapers, magazines, and
documents which are the outputs of typewriters, printers or plotters. OCR systems of machineprinted documents can be classified into [5, 6]:
• Mono-font OCR systems that deals with documents written with one specific font,
• Multi-font OCR systems that handles a subset of the existing fonts (recognition of more than
one font),
• Omni-font OCR systems that allows the recognition of characters in any font.
Today many types of OCR software available in the markets like: Desktop OCR, Server OCR,
Web OCR etc. Accuracy rate of any OCR tool varies from 71% to 98% [7].
2. PROBLEM DEFINITION
In modern society, we rely heavily on computers to process huge volumes of data. Related to
this and for economic reasons or business requirements, there is a great demand for quickly
converting the printed information in a document into an edited text in the computer. Often these
data exist on paper and they have to be typed into the computer by human operators. Such timeconsuming and error-prone processes have been lightened by the invention of OCR systems [1].
Unfortunately, OCR systems are still erroneous and inaccurate, especially when the source document
is of low printing quality [2]. Therefore, accuracy of these systems can be dependent on text
preprocessing and segmentation algorithms. Sometimes it is difficult to retrieve text from the image
because of different size, style, orientation, complex background of image … etc which produce
misspellings in the recognized text [2, 7].
OCR technology allows machine to recognize the text automatically in the same way as the
combination of eye and mind of human body. In development of computerized OCR system, few
problems can occur [7]:
• There is very little visible difference between some letters and digits for computers to
understand. For example it might be difficult for the computer to differentiate between digit
'0' and letter 'o' / 'O',
• It might be very difficult to extract text, which is embedded in very dark background or
printed on other words or graphics.

49
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

3. AIM OF THE WORK
This paper aims to build an OCR Multi-font system which converts the printed English texts
in a paper document (optical patterns exist in a digital image) into an edited text (its corresponding
alphanumeric form or other symbols) in the computer in order to:
• eliminate the time-consuming spent by human operators to re-type the documents (the huge
volumes of data),
• reduce the possible errors occurred in the typing process,
• save money (by cancelling the need of human typists),
• preserve space needed for paper books.
4. THE PROPOSED OCR SYSTEM
The following block diagram, shown in Figure (1), illustrates the proposed OCR system model.

Figure (1): Block-Diagram of the Proposed OCR System Model
The input scanned image text is passed through a sequence of preprocessing steps (noise
removal, foreground/ background separation, normalization, and binarization) prior to characters
segmentation phase. Then, feature extraction methods (Discrete Cosine Transform (DCT) followed
by Zigzag process, or Wavelet Transform (WT)) are applied to the segmented characters. The
obtained feature-set is either stored in database as templates or references during the training phase
when building database (DB) of characters features-set or is compared directly during the testing
phase to those DB references in a pattern matching stage. Finally, decision rule is applied to produce
recognition results beside the best matched characters.

50
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

4.1 The Input Scanned Image
The input documents used for training and testing are scanned and digitized by a page
scanner at 300 dpi resolution connected to a computer system and saved in BMP format of 256 graylevels. A number of scanned images that are used as inputs to our OCR system model are shown in
Figure (2) below:

Figure (2): 3–Samples of Scanned Images
4.2 Document Image Preprocessing
Digital Images are generally corrupted by noise during the process of acquisition and
transmission. This noise degrades the quality of digital image which produces several tiny dots
scattered due to uneven gray-scale intensities which causes poor recognition rates. Consequently the
performance of any system manipulating these images is also decreased. Therefore, removal of noise
in document images corrupted by Gaussian and Impulse noises before OCR is important to guarantee
better accuracy of characters recognition [1, 8].
Thus, image enhancement techniques adopted in this system model in the following sequence
are employed prior to segmentation in order to simplify the process of characters segmentation [9,
10]:
a) Noise Removal: is applied to the scanned document images for two primary purposes: to
eliminate the noise and to give an image a softer effect. The spatial convolution mask of
Gaussian filter used for low-pass filtering is shown in Figure (3).

Figure (3): Gauss core – weighted on distance (Gaussian filter)
Gaussian filter smoothen the image to match the pixelsnearby in a way that no point in the image
differ from its surroundings to agreater extent. Image smoothing is accomplished in the spatial
domain toremove errors, incorrect data and simplify the acquisition process of patterns
[8, 10, 11].
51
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

b) Foreground/ Background Separation:is the process of separating the foreground regions (the
area of interest containing the printed-text) in the image from the background regions (the
useless area outside the borders of printed-text). The background regions generally exhibit a very
low gray-scale variance value, whereas the foreground regions have a very high variance. Hence,
a method based on variance thresholding can be used to perform the separation. Firstly, the
image is divided into blocks and thegray-scale variance is calculated for each block in the image.
If the variance is less than the global threshold, then the block is assigned to be a background
region; otherwise, it is assigned to be part of the foreground. The gray-level variance for a block
of size W × W is defined as [9, 11]:

1
V (k ) =
W 2

W −1 W −1

∑ ∑ ( I (i,
i=0

j) − M (k )

)2

…… (1)

j=0

where V(k) is the variance for kth block, I(i, j) is the gray-level value at pixel(i, j), and M(k) is the
mean gray-level value for the kth block.
c) Normalization: is utilized to reduce the effect of non-uniform intensities and improving image
quality by stretching its histogram. To be able to normalize an image the area which is to
normalize within has tobe known. Thus it is necessary to find the highest and the lowest pixel
value ofthe current image. Every pixel is then evenly spread out along the scale by thefollowing
equation [10, 11]:

N (i, j ) =

I ( i , j ) − I min
× M
I max − I min

…… (2)

whereI(i, j) is the gray-level value at pixel (i, j), Imin is the smallest gray-level value found in the
image, Imax is the largest gray-level value found in the image, Mrepresents the new maximum
gray-level value of the scale (i.e., M=255), and N(i, j) represent the normalized gray-level value
at pixel (i, j).
d) Binarization:Is the process of turning a gray scale image into a binary image (only two levels of
interest 0 and 1) in order to improve the contrast, and consequently facilitates the feature
extraction process.It is impossible to find a working global threshold value that can be used
efficiently on every image because of the variations among the scanned images.Therefore
algorithms to find the optimal value, based on localized thresholds, must be applied separately
on each image to get a functional binarization. The image is partitioned into smaller blocks and
threshold values are then calculated for each of these blocks. This enables adaptations that are
not possible with global calculations. Localized thresholds demand a lot more calculations but
mostly compensate it with better results [9, 11].The local mean threshold for kthblock of size W
× W is computed below:

1 W −1 W −1
Local Mean ( k ) = 2 ∑ ∑ Block (i , j )
W i =0 j =0

…… (3)

whereBlock (i, j) is the gray-level value at pixel (i, j).If the pixel value is lower than the threshold
then the pixel is assigned to be part of the Printed-text; otherwise, it is assigned to be part of
Background.
52
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

Figure (4) illustrates the effect of applying the document image preprocessing techniques to a
scanned document sample.

Figure (4): The Effects of the proposed preprocessing steps on a scanned document image
sample
From Figure (4.c), it is noticeable that background regions are specified by black color,
which is only to indicate the effect of this process despite the fact that in practical application the
background is represented by white color.
4.3 Characters Segmentation
Segmentation is an important phase and the accuracy of any OCR heavily depends on it,
where incorrect segmentation leads to reduction in recognition accuracy. Segmentation is the process
thatdivides the whole document into smaller components, which include [6, 12]:
•
•

Line,
Word, and Character segmentation

The procedure adopted in this work for analyzing imagesto detect characters, as shown in
Figure (5), is listed in the following sequence:
Step1: Perform "Row Scan" (Line segmentation) to find the number of lines and boundaries of
each line in any input document image within which the detection can proceed.

53
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

Algorithm1
a)
b)
c)
d)
e)
f)
g)
h)
i)

j)

Get image data from the processed image.
Initialize lines boundaries (T: Top, B: Bottom) to -1.
Perform row scan (from 1st to last row) for pixels value= 0 (i.e, black).
Set the number of lines to 0.
If black pixel is detected then register T as the top of the current line and move the pointer to
the next row & 1st column. Otherwise, continue to the next pixel (from left to right).
If black pixel is found and T <> -1 then update B as the bottom of the current line and move
the pointer to the next row & 1st column.
If no black pixel is found in the row and T <> -1 then increment the number of lines by 1.
Start below the bottom of the last line found and repeat steps e) – g) to detect subsequent lines
(Stop when the bottom of image is reached).
Process dot "." problem spacing found in (i and j) characters by merging lines below certain
space threshold according to the font type & size and decrement the number of lines by 1 (if
such case occurred).
Print out the number of lines detected and draw line boundaries on the processed image for
each detected lines (for: T–1 and B+1).

Step2: Perform "Column Scan" (orthogonally from 1st to last column) only for detected lines.
Thus, detecting characters in an image does not necessarily involve scanning the whole image all
over again.
Algorithm2
a)
b)
c)
d)

e)
f)
g)
h)

Initialize characters boundaries (L: Left, R: Right) to -1.
Perform column scan (from 1st to last column) for pixels value= 0 (i.e, black).
Set the number of characters to 0.
If black pixel is detected then register L as the left of the current character and move the
pointer to the next column & 1st row. Otherwise, continue to the next pixel (from top to
bottom).
If black pixel is found and L <> -1 then update R as the right of the current character and
move the pointer to the next column & 1st row.
If no black pixel is found in the column and L <> -1 then increment the number of characters
by 1.
Scan up to the right of the character found and repeat steps d) – f) to detect subsequent
characters (Stop when the right-end of the last line is reached).
Print out the number of characters detected and draw line boundaries on the image for each
detected characters (for:L–1 andR+1).

Step3: Perform "Row Scan"once more on the results obtained from the previous step in order to
detect the actual character (top & bottom) boundaries.
Step4: Bitmap imagesare created on the hard-disk for each segmented character relative to its
boundaries; and its header information is generated from the original scanned image header besides
updating dimensions.
54
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

Step 2
Character bounds
Left & Right

Step 1

Line Top

1

Line Bottom

2

Step 3
Character Top
Character Bottom

3
4

Figure (5): Lines and Characters boundary detection for a scanned image sample
From the above figure, it is obvious that the detected lines-bound (top & bottom) stated by
the "red-color" might not necessarily be the actual bounds for the characters in the same line because
the characters have different heights. Hence, a confirmation of top and bottom boundaries for each
character is needed as stated by the "green-color". The "blue-color" illustrates the detected
characters-bound (left & right) for different characters widths.
4.4 Database Construction
In general, any OCR system depends on training samples as input data. In this work, database
of (91) samples were collected from the data sets shown in Table (1) below for different font types
("Arial", "Calibri", "Courier New", "Lucida Sans" and "Times New Roman") and for different font
sizes (8, 10, 12, 14, 16, 18, 20) to produce (3185) samples.

No.
1
2
3
4

Table (1): Data-Set Samples
No. of
Data Sets
Samples
Samples
Digits
Capital English
Letters
Small English
Letters
Some common
ASCII Symbols

10

0 1 2 … 9

26

'A' 'B' 'C' … 'Z'

26

'a' 'b' 'c' … 'z'

29

.,;'":[]-+ */=(){}
<> ! @ # $ % ^ & ?  _

4.5 Feature Extraction
Feature extraction is part of the data reduction process by forming a new “smaller” set of
features from the original feature set of the patterns. This can be done by extracting some numerical
55
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

measurements from raw input patterns. Image features are of major importance in the isolation of
regions of common property within an image[9, 10, 13]. In this work, two-sets of features were
extracted from the segmented characters either by the use of Discrete Cosine Transform (DCT) or
the spectral properties of Wavelet transform.
a) Discrete Cosine Transform (DCT)
DCT has become a standard method for many image processing& video compression
algorithms. The two-dimensional DCT can be computed using the one-dimensional DCT
horizontally (row-wise) and then vertically (column-wise) across the image because DCT is a
function that separates the image into frequencies with large variance. The two-dimensional Forward
DCT (2D FDCT) coefficients F(u, v)of M x N block of image pixels f(x , y)are formulated as [11, 13,
14, 15]:
…… (4)

…… (5)

where C(k) is the normalization constant, u = 0,1,. . . , N – 1 and v = 0,1,. . . , M – 1.
The DCT coefficients (Di,j and i, j = 0,1, … 7) of the corresponding image block of size 8×8,
as example, are then ordered in a particular irregular sequence as shown in Figure (6). This irregular
ordering of the coefficients is called Zig-zag ordering.

Figure (6): Zig-zag ordering of DCT Coefficients
The above sequence is broken into runs of nonzero (the early coefficients that contains the
important "low-frequency" image information) and zero coefficients (the later coefficients in a block
that contains the less-important "high-frequency" image information) [14, 15].Therefore, the final
DCT feature-set considered in this work is generated only from the number of significant (nonzero)
coefficients denoted by N in Table (2) that starts from D0,0.
b) Wavelet Transform
The wavelet transform has been found very useful for the time-scale representation and has
been widely used in signal processing and computer vision. The Wavelet transform is a multiresolution technique that cut up data into different frequency components, and then analyzes each
component with a resolution matched to its scale. The forward and inverse continuous wavelet
56
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

transform of x(t)"the signal to be analyzed" with respect to the basis function or wavelet ψ j , k (t) at
scale j (j>0) and time delay k is written as follows [16, 17, 18]:

Forward CWT: W( j,k ) =

Inverse CWT :
where

∫ x(t)

ψ

(t) dt

…… (6)

x(t) = ∫ ∫ W(j,k) ψ j, k (t) dk dj

…… (7)

j,k

k j

ψ j,k (t) = 1 ψ( t − k )
j

j

and

ψ(t) is the mother wavelet …… (8)

This multiresolution can also be obtained using filter banks, resulting in the Discrete Wavelet
Transform (DWT) that are well suited to the digital computer because there are no derivatives or
integrals, just multiplication and addition operations which correspond to the mathematical
convolution operation. The procedure starts with passing the signal (sequence) through a half-band
digital low-pass & high-pass filters. The DWT is computed by successive low pass and high pass
filtering of the discrete time-domain signal X[n], as shown in Figure (7), and each result is downsampled by two (↓2) where the low-pass filter is denoted by G0 while the high-pass filter is denoted
by H0. At each level, the high-pass filter produces detail information d[n] while the low-pass filter
associated with scaling function produces coarse approximations a[n].The DWT of the original
signal is then obtained by concatenating all the coefficients, a[n] and d[n], starting from the last level
of decomposition [10, 17, 18].

Figure (7): Three Levels DWT Decomposition Tree
After the conversion of input image from its lowest-level of pixel data in spatial domain I(x,
y) into higher-level representation of wavelet coefficients W(x, y), a set of wavelet features (the
energy of each band as stated in Eq.(9))can be extracted by recursively decomposing sub-images in
the low frequency channels as shown in Figure (8).The number of wavelet features for the 1st level is
4, and each progressing in wavelet level will correspond an increasing in features length by 3.
…… (9)

57
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

Figure (8): Three-Level Wavelet Multiresolution Decomposition Image
Wavelet transform breaks an image down into four sub-sampled images. The results consist
of one image that has been high pass filtered in both horizontal and vertical directions (HH), one that
has been high pass filtered in the vertical and low pass filtered in the horizontal (LH), one that has
been low passed in the vertical and high passed in the horizontal (HL), and one that has been low
pass filtered in both directions(LL).
Numerous filters can be used to implement the wavelet transform. Daubechies (D4) wavelet
is one of the most commonly used due to its efficiency. The Daubechies basis vectors are [10, 16,
18]:
…… (10)
…… (11)
4.6 Pattern Matching
The resulting test template, which is an N-dimensional feature vector, is compared against the
stored reference templates to find the closest match. The process is to find which unknown class
matches a predefined class or classes. For the OCR task, the unknown character is compared to all
references in the database. This comparison can be done through Euclidean (E.D.) distance measure,
shown below [9, 10]:
…… (12)
N

E .D . =

∑

i =1

(a i − b i )2

where A and B are two vectors, such that A=[a1a2… aN]andB=[b1 b2… bN].
In our approach the minimum distance classifier is used to measure the difference between
the two patterns (feature vectors). This classifier assigns the unknown pattern to the nearest
predefined pattern. The bigger distance between the two vectors, is the greater difference [9, 10].
5. EXPERIMENTAL RESULTS
OCR accuracy is definedas the ratio of correct recognized characters to the total number of
characters (samples) tested,as shown by Eq. (13):
58
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

OCR Accuracy =

No. of Correctly Re cognized Characters
× 100 %
Total No. of Characters Tested

..… (13)

A number of experiments and test conditions were accomplished on (13650) samples to
measure the performance of the proposed OCR system on various types of document images of
different dimensions, font types and font sizes. As a result, the database size of training samples is
computed as follows:
No . of Training Samples = No . of Samples × No . of Fonts × No . of Font Sizes

..… (14)

= 91× 5 × 7 = 3185 Samples

A more appropriate comparison can be made if both DCT & Wavelet methods are measured
under identical conditions. Based on the results shown in Table (2), one can deduce that all wavelet
features produce better recognition rates 96% (from 92 – 98%) than DCT features 92% (from 87 –
95%). Different number of DCT coefficients (N) and wavelet decomposition levels values are
examined according to the recognition rates. It is clearly indicated that two decomposition levels are
most appropriate for wavelet feature vector construction, whereas 10-DCT coefficients is the suitable
number of DCT features (N=10).
Table (2): OCR Accuracy for different Testing images using DCT &Wavelet features
Test Image
File

Font
Type

No. of
Chars

DCT

Wavelet Transform

Test1.bmp
Test2.bmp
Test3.bmp
Arial
Test4.bmp
Test5.bmp
Test6.bmp
Test7.bmp
Test8.bmp Calibri
Test9.bmp
Test10.bmp
Test11.bmp
Test12.bmp
Courier
Test13.bmp
New
Test14.bmp
Test15.bmp
Test16.bmp
Test17.bmp
Lucida
Test18.bmp
Sans
Test19.bmp
Test20.bmp
Test21.bmp
Test22.bmp Times
Test23.bmp
New
Test24.bmp Roman
Test25.bmp

525
581
791
371
462
480
633
415
692
510
395
704
620
557
454
664
325
653
341
747
720
373
475
525
637

N=5
77.52
79.00
81.29
75.74
71.65
76.46
79.78
78.31
79.05
77.06
75.70
81.82
79.35
79.35
75.11
80.72
72.92
78.56
68.62
79.65
76.11
71.05
73.68
79.05
78.02

N=10
91.62
93.46
90.90
94.88
87.88
92.29
93.84
95.66
90.75
92.75
92.41
95.60
92.58
93.36
91.63
93.67
94.46
92.04
89.15
91.30
90.28
87.13
89.05
93.14
90.74

N=15
85.90
87.44
85.59
83.83
78.35
82.29
85.47
84.10
83.38
84.31
82.28
86.36
83.23
84.02
81.28
86.30
82.15
82.54
77.71
82.33
79.44
80.43
79.79
82.86
81.48

Level1
71.43
74.87
80.28
64.15
69.05
75.63
77.09
68.43
75.29
79.02
67.34
73.86
75.97
74.33
68.06
75.75
68.92
76.11
69.21
78.71
77.64
72.12
68.63
74.10
78.96

Level2
94.29
96.73
94.82
97.30
93.51
95.00
95.10
96.63
96.68
97.25
96.46
97.30
96.61
98.20
96.92
95.63
92.62
94.49
95.60
94.38
96.81
98.93
95.37
96.95
97.02

Level3
80.95
87.78
85.59
80.32
83.77
83.75
80.73
83.37
85.55
84.90
81.27
84.52
83.71
89.23
82.60
84.19
79.38
82.24
78.01
83.53
85.00
86.33
80.63
85.52
86.81

Total

13650

77.64

92.05

83.16

74.25

96.01

83.89

59
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

A good feature set has a major importance in the isolation of regions of common property
within an image, and it should represent characteristic of a class that help distinguish it from other
classes, as shown in Figure (9).The proportion of (‫ ) ܹ⁄ ܪ‬was taken in our consideration for all text
components (characters, digits, and symbols) for two reasons:
• it is invariant to different font sizes of the same font type,
• to speed-up the recognition process because the matching processes are limited only to the
text components in the same class.
Consequently, an improvement overall recognition rates (about 3%) are obtained after classifying
characters according to the addition of a new discriminating feature (the proportion of Height to
Width) that produce 99% for wavelet and 95% for DCT.

Figure (9): The Proportion of (۶⁄‫ )܅‬Feature for the character sample 'E' in different Font
types & sizes
6. CONCLUSIONS
A multi-fonts English texts OCR system for 3185 training samples and 13650 testing samples
is presented that relies on DCT and wavelet features. Image enhancement techniques (noise removal,
foreground/ background separation, normalization and binarization) are adopted in this work prior to
segmentation in order to improve the recognition rates and simplify the process of characters
segmentation.
It is found that wavelet method is appropriate for feature vector construction where all the
recognition rates (96%) outperform the DCT based recognition method (92%). To enhance the
recognition rates further and speeds up the recognition process, text-components (characters, digits,
and symbols) are classified according to the proportion of (‫ ) ܹ⁄ ܪ‬feature that produce 99% accuracy
for wavelet based method because it help distinguish a class from other classes which is invariant to
different font sizes of the same font type.
REFERENCES
[1]
[2]

Mohamed Cheriet, NawwafKharma, Cheng-Lin Liu, ChingY.Suen, "Character Recognition
Systems: A Guide for Students and Practitioners", John Wiley & Sons, Inc., Canada, 2007.
Youssef Bassil, Mohammad Alwani, "OCR Context-Sensitive Error Correction Based on
Google Web 1T 5-Gram Data Set", American Journal of Scientific Research, ISSN 1450-223X,
Issue. 50, February 2012.
60
International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print),
ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME

[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]

[21]

R. JagadeeshKannan, R. Prabhakar, "An Improved Handwritten Tamil Character
Recognition System using Octal Graph", Journal of Computer Science 4 (7): 509-516, ISSN
1549-3636, 2008.
R. C. Tripathi, Vijay Kumar, "Character Recognition: A Neural Network Approach",
National Conference on Advancement Technologies – Information Systems & Computer
Networks (ISCON) published in IJCA®, 2012.
Nafiz Arica, "An Off-Line Character Recognition System for Free Style Handwriting",
M.Sc. thesis submitted to the department of computer engineering, the graduate school of natural
and applied sciences of the Middle East Technical University, 1998.
Ibrahim Abuhaiba, "Arabic Font Recognition Based on Templates", the International Arab
Journal of Information Technology, Vol. 1, No. 0, July 2003.
Chirag Patel, Atul Patel, Dharmendra Patel, "Optical Character Recognition by Open Source
OCR Tool Tesseract: A Case Study", International Journal of Computer Applications (0975 –
8887) Vol. 55 – No.10, October 2012.
C. Patvardhan, A. K. Verma, C. V. Lakshmi, "Denoising of Document Images using Discrete
Curvelet Transform for OCR Applications", International Journal of Computer Applications
(0975 – 8887), Vol. 55 – No.10, October 2012.
Rafael C. Gonzalez, Richard E. Woods, “Digital Image Processing”, Second Edition, PrenticeHall, Inc., New Jersey, U.S.A., 2007.
S. E. Umbaugh, “Computer Vision and Image Processing”, Prentice-Hall, Inc., U.S.A., 1998.
TinkuAcharya, Ajoy K. Ray, "Image Processing: Principles and Applications", John Wiley &
Sons, Inc., New Jersey, U.S.A., 2005.
Ashu Kumar, Simpel Rani Jindal, "Segmentation of handwritten Gurmukhi text into lines",
International Conference on Recent Advances and Future Trends in Information Technology
(iRAFIT) published in IJCA®, 2012.
William K. Pratt, "Digital Image Processing", Fourth Edition, John Wiley & Sons, Inc., New
Jersey, U.S.A., 2007.
Milan Sonka, Vaclav Hlavac, Roger Boyle, "Image Processing, Analysis and Machine Vision",
Third International Student Edition, Thomson Corporation, U.S.A., 2008.
David Salomon, "Data Compression: The Complete Reference", Fourth Edition, SpringerVerlag London Limited 2007.
C. S. Burrus, R. A. Gopinath and H. Guo, “Introduction to Wavelets and Wavelet
Transforms”, Prentice-Hall, Inc., U.S.A, 1998.
M. Kociołek, A. Materka, M. Strzelecki P. Szczypiński, "Discrete Wavelet Transform –
Derived Features for Digital Image Texture Analysis", Proc. of International Conference on
Signals and Electronic Systems, pp. 163-168, Poland, 2001.
V. Jeengar, S.N. Omkar, A. Singh, "A Review Comparison of Wavelet and Cosine Image
Transforms", I.J.Image, Graphics and Signal Processing, DOI, 2012.
M. M. Kodabagi, S. A. Angadi and Chetana. R. Shivanagi, “Character Recognition of Kannada
Text in Scene Images using Neural Network”, International Journal of Graphics and Multimedia
(IJGM), Volume 4, Issue 1, 2013, pp. 9 - 19, ISSN Print: 0976 – 6448, ISSN Online: 0976 –6456.
Dr. Mustafa Dhiaa Al-Hassani and Dr. Abdulkareem A. Kadhim, “Design a Text-Prompt Speaker
Recognition System using Lpc-Derived Features”, International Journal of Information
Technology and Management Information Systems (IJITMIS), Volume 4, Issue 3, 2013,
pp. 68 - 84, ISSN Print: 0976 – 6405, ISSN Online: 0976 – 6413.
Dr. Mustafa Dhiaa Al-Hassani, Dr. Abdulkareem A. Kadhim and Dr. Venus W. Samawi,
“Fingerprint Identification Technique Based on Wavelet-Bands Selection Features (WBSF)”,
International Journal of Computer Engineering & Technology (IJCET), Volume 4, Issue 3, 2013,
pp. 308 - 323, ISSN Print: 0976 – 6367, ISSN Online: 0976 – 6375.

61

Mais conteúdo relacionado

Mais procurados

Optical character recognition (ocr) ppt
Optical character recognition (ocr) pptOptical character recognition (ocr) ppt
Optical character recognition (ocr) pptDeijee Kalita
 
Final Report on Optical Character Recognition
Final Report on Optical Character Recognition Final Report on Optical Character Recognition
Final Report on Optical Character Recognition Vidyut Singhania
 
Handwritten Text Recognition and Digital Text Conversion
Handwritten Text Recognition and Digital Text ConversionHandwritten Text Recognition and Digital Text Conversion
Handwritten Text Recognition and Digital Text Conversionijtsrd
 
OCR vs. Urjanet
OCR vs. UrjanetOCR vs. Urjanet
OCR vs. UrjanetUrjanet
 
Project report of OCR Recognition
Project report of OCR RecognitionProject report of OCR Recognition
Project report of OCR RecognitionBharat Kalia
 
Optical Character Recognition Using Python
Optical Character Recognition Using PythonOptical Character Recognition Using Python
Optical Character Recognition Using PythonYogeshIJTSRD
 
Optical Character Recognition (OCR) System
Optical Character Recognition (OCR) SystemOptical Character Recognition (OCR) System
Optical Character Recognition (OCR) Systemiosrjce
 
OCR speech using Labview
OCR speech using LabviewOCR speech using Labview
OCR speech using LabviewBharat Thakur
 
A STUDY ON OPTICAL CHARACTER RECOGNITION TECHNIQUES
A STUDY ON OPTICAL CHARACTER RECOGNITION TECHNIQUESA STUDY ON OPTICAL CHARACTER RECOGNITION TECHNIQUES
A STUDY ON OPTICAL CHARACTER RECOGNITION TECHNIQUESijcsitcejournal
 
Optical Character Recognition (OCR)
Optical Character Recognition (OCR)Optical Character Recognition (OCR)
Optical Character Recognition (OCR)Vidyut Singhania
 
Handwriting Recognition
Handwriting RecognitionHandwriting Recognition
Handwriting RecognitionBindu Karki
 
Praveen
PraveenPraveen
Praveenrjmktg
 
Handwritten Character Recognition: A Comprehensive Review on Geometrical Anal...
Handwritten Character Recognition: A Comprehensive Review on Geometrical Anal...Handwritten Character Recognition: A Comprehensive Review on Geometrical Anal...
Handwritten Character Recognition: A Comprehensive Review on Geometrical Anal...iosrjce
 
Optical Character Recognition System for Urdu (Naskh Font)Using Pattern Match...
Optical Character Recognition System for Urdu (Naskh Font)Using Pattern Match...Optical Character Recognition System for Urdu (Naskh Font)Using Pattern Match...
Optical Character Recognition System for Urdu (Naskh Font)Using Pattern Match...CSCJournals
 
Handwriting Recognition Using Deep Learning and Computer Version
Handwriting Recognition Using Deep Learning and Computer VersionHandwriting Recognition Using Deep Learning and Computer Version
Handwriting Recognition Using Deep Learning and Computer VersionNaiyan Noor
 

Mais procurados (20)

Optical character recognition (ocr) ppt
Optical character recognition (ocr) pptOptical character recognition (ocr) ppt
Optical character recognition (ocr) ppt
 
Final Report on Optical Character Recognition
Final Report on Optical Character Recognition Final Report on Optical Character Recognition
Final Report on Optical Character Recognition
 
Handwritten Text Recognition and Digital Text Conversion
Handwritten Text Recognition and Digital Text ConversionHandwritten Text Recognition and Digital Text Conversion
Handwritten Text Recognition and Digital Text Conversion
 
Ocr abstract
Ocr abstractOcr abstract
Ocr abstract
 
OCR vs. Urjanet
OCR vs. UrjanetOCR vs. Urjanet
OCR vs. Urjanet
 
Project report of OCR Recognition
Project report of OCR RecognitionProject report of OCR Recognition
Project report of OCR Recognition
 
Basics of-optical-character-recognition
Basics of-optical-character-recognitionBasics of-optical-character-recognition
Basics of-optical-character-recognition
 
Optical Character Recognition Using Python
Optical Character Recognition Using PythonOptical Character Recognition Using Python
Optical Character Recognition Using Python
 
Optical Character Recognition (OCR) System
Optical Character Recognition (OCR) SystemOptical Character Recognition (OCR) System
Optical Character Recognition (OCR) System
 
OCR speech using Labview
OCR speech using LabviewOCR speech using Labview
OCR speech using Labview
 
A STUDY ON OPTICAL CHARACTER RECOGNITION TECHNIQUES
A STUDY ON OPTICAL CHARACTER RECOGNITION TECHNIQUESA STUDY ON OPTICAL CHARACTER RECOGNITION TECHNIQUES
A STUDY ON OPTICAL CHARACTER RECOGNITION TECHNIQUES
 
En31919926
En31919926En31919926
En31919926
 
Optical Character Recognition (OCR)
Optical Character Recognition (OCR)Optical Character Recognition (OCR)
Optical Character Recognition (OCR)
 
OCR2
OCR2OCR2
OCR2
 
Handwriting Recognition
Handwriting RecognitionHandwriting Recognition
Handwriting Recognition
 
OCR Text Extraction
OCR Text ExtractionOCR Text Extraction
OCR Text Extraction
 
Praveen
PraveenPraveen
Praveen
 
Handwritten Character Recognition: A Comprehensive Review on Geometrical Anal...
Handwritten Character Recognition: A Comprehensive Review on Geometrical Anal...Handwritten Character Recognition: A Comprehensive Review on Geometrical Anal...
Handwritten Character Recognition: A Comprehensive Review on Geometrical Anal...
 
Optical Character Recognition System for Urdu (Naskh Font)Using Pattern Match...
Optical Character Recognition System for Urdu (Naskh Font)Using Pattern Match...Optical Character Recognition System for Urdu (Naskh Font)Using Pattern Match...
Optical Character Recognition System for Urdu (Naskh Font)Using Pattern Match...
 
Handwriting Recognition Using Deep Learning and Computer Version
Handwriting Recognition Using Deep Learning and Computer VersionHandwriting Recognition Using Deep Learning and Computer Version
Handwriting Recognition Using Deep Learning and Computer Version
 

Destaque

A pattern diversity compact mimo antenna array design for wlan
A pattern diversity compact mimo antenna array design for wlanA pattern diversity compact mimo antenna array design for wlan
A pattern diversity compact mimo antenna array design for wlanIAEME Publication
 
An empirical large signal model for rf ldmosfet transistors
An empirical large signal model for rf ldmosfet transistorsAn empirical large signal model for rf ldmosfet transistors
An empirical large signal model for rf ldmosfet transistorsIAEME Publication
 
An empirical large signal model for rf ldmosfet transistors
An empirical large signal model for rf ldmosfet transistorsAn empirical large signal model for rf ldmosfet transistors
An empirical large signal model for rf ldmosfet transistorsIAEME Publication
 

Destaque (9)

A pattern diversity compact mimo antenna array design for wlan
A pattern diversity compact mimo antenna array design for wlanA pattern diversity compact mimo antenna array design for wlan
A pattern diversity compact mimo antenna array design for wlan
 
10220130403003
1022013040300310220130403003
10220130403003
 
50120130406006
5012013040600650120130406006
50120130406006
 
20320140504005
2032014050400520320140504005
20320140504005
 
20320130406003 2-3
20320130406003 2-320320130406003 2-3
20320130406003 2-3
 
An empirical large signal model for rf ldmosfet transistors
An empirical large signal model for rf ldmosfet transistorsAn empirical large signal model for rf ldmosfet transistors
An empirical large signal model for rf ldmosfet transistors
 
20120140504023
2012014050402320120140504023
20120140504023
 
10120140504017
1012014050401710120140504017
10120140504017
 
An empirical large signal model for rf ldmosfet transistors
An empirical large signal model for rf ldmosfet transistorsAn empirical large signal model for rf ldmosfet transistors
An empirical large signal model for rf ldmosfet transistors
 

Semelhante a 50120130406005

optical character recognition system
optical character recognition systemoptical character recognition system
optical character recognition systemVijay Apurva
 
Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Editor IJARCET
 
Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Editor IJARCET
 
How to create a corpus of machine-readable texts: challenges and solutions
How to create a corpus of machine-readable texts: challenges and solutionsHow to create a corpus of machine-readable texts: challenges and solutions
How to create a corpus of machine-readable texts: challenges and solutionsMonika Renate Barget
 
Input and output devices
Input and output devicesInput and output devices
Input and output devicesShouaQureshi
 
IRJET- Intelligent Character Recognition of Handwritten Characters using ...
IRJET-  	  Intelligent Character Recognition of Handwritten Characters using ...IRJET-  	  Intelligent Character Recognition of Handwritten Characters using ...
IRJET- Intelligent Character Recognition of Handwritten Characters using ...IRJET Journal
 
Correcting optical character recognition result via a novel approach
Correcting optical character recognition result via a novel approachCorrecting optical character recognition result via a novel approach
Correcting optical character recognition result via a novel approachIJICTJOURNAL
 
Opticalcharacter recognition
Opticalcharacter recognition Opticalcharacter recognition
Opticalcharacter recognition Shobhit Saxena
 
BLOB DETECTION TECHNIQUE USING IMAGE PROCESSING FOR IDENTIFICATION OF MACHINE...
BLOB DETECTION TECHNIQUE USING IMAGE PROCESSING FOR IDENTIFICATION OF MACHINE...BLOB DETECTION TECHNIQUE USING IMAGE PROCESSING FOR IDENTIFICATION OF MACHINE...
BLOB DETECTION TECHNIQUE USING IMAGE PROCESSING FOR IDENTIFICATION OF MACHINE...ijiert bestjournal
 
IRJET- Offline Transcription using AI
IRJET-  	  Offline Transcription using AIIRJET-  	  Offline Transcription using AI
IRJET- Offline Transcription using AIIRJET Journal
 
OCR 's Functions
OCR 's FunctionsOCR 's Functions
OCR 's Functionsprithvi764
 
IRJET- Optical Character Recognition using Image Processing
IRJET-  	  Optical Character Recognition using Image ProcessingIRJET-  	  Optical Character Recognition using Image Processing
IRJET- Optical Character Recognition using Image ProcessingIRJET Journal
 

Semelhante a 50120130406005 (20)

optical character recognition system
optical character recognition systemoptical character recognition system
optical character recognition system
 
Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015
 
Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015
 
Z04405149151
Z04405149151Z04405149151
Z04405149151
 
D017222226
D017222226D017222226
D017222226
 
How to create a corpus of machine-readable texts: challenges and solutions
How to create a corpus of machine-readable texts: challenges and solutionsHow to create a corpus of machine-readable texts: challenges and solutions
How to create a corpus of machine-readable texts: challenges and solutions
 
Input and output devices
Input and output devicesInput and output devices
Input and output devices
 
INPUT AND OUTPUT DEVICES
INPUT AND OUTPUT DEVICESINPUT AND OUTPUT DEVICES
INPUT AND OUTPUT DEVICES
 
IRJET- Intelligent Character Recognition of Handwritten Characters using ...
IRJET-  	  Intelligent Character Recognition of Handwritten Characters using ...IRJET-  	  Intelligent Character Recognition of Handwritten Characters using ...
IRJET- Intelligent Character Recognition of Handwritten Characters using ...
 
05a
05a05a
05a
 
Paper based interaction
Paper based interactionPaper based interaction
Paper based interaction
 
O45018291
O45018291O45018291
O45018291
 
Correcting optical character recognition result via a novel approach
Correcting optical character recognition result via a novel approachCorrecting optical character recognition result via a novel approach
Correcting optical character recognition result via a novel approach
 
CRC Final Report
CRC Final ReportCRC Final Report
CRC Final Report
 
A12REVIEW.pptx
A12REVIEW.pptxA12REVIEW.pptx
A12REVIEW.pptx
 
Opticalcharacter recognition
Opticalcharacter recognition Opticalcharacter recognition
Opticalcharacter recognition
 
BLOB DETECTION TECHNIQUE USING IMAGE PROCESSING FOR IDENTIFICATION OF MACHINE...
BLOB DETECTION TECHNIQUE USING IMAGE PROCESSING FOR IDENTIFICATION OF MACHINE...BLOB DETECTION TECHNIQUE USING IMAGE PROCESSING FOR IDENTIFICATION OF MACHINE...
BLOB DETECTION TECHNIQUE USING IMAGE PROCESSING FOR IDENTIFICATION OF MACHINE...
 
IRJET- Offline Transcription using AI
IRJET-  	  Offline Transcription using AIIRJET-  	  Offline Transcription using AI
IRJET- Offline Transcription using AI
 
OCR 's Functions
OCR 's FunctionsOCR 's Functions
OCR 's Functions
 
IRJET- Optical Character Recognition using Image Processing
IRJET-  	  Optical Character Recognition using Image ProcessingIRJET-  	  Optical Character Recognition using Image Processing
IRJET- Optical Character Recognition using Image Processing
 

Mais de IAEME Publication

IAEME_Publication_Call_for_Paper_September_2022.pdf
IAEME_Publication_Call_for_Paper_September_2022.pdfIAEME_Publication_Call_for_Paper_September_2022.pdf
IAEME_Publication_Call_for_Paper_September_2022.pdfIAEME Publication
 
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...IAEME Publication
 
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURSA STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURSIAEME Publication
 
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURSBROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURSIAEME Publication
 
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONSDETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONSIAEME Publication
 
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONSANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONSIAEME Publication
 
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINOVOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINOIAEME Publication
 
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...IAEME Publication
 
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMYVISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMYIAEME Publication
 
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...IAEME Publication
 
GANDHI ON NON-VIOLENT POLICE
GANDHI ON NON-VIOLENT POLICEGANDHI ON NON-VIOLENT POLICE
GANDHI ON NON-VIOLENT POLICEIAEME Publication
 
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...IAEME Publication
 
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...IAEME Publication
 
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...IAEME Publication
 
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...IAEME Publication
 
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...IAEME Publication
 
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...IAEME Publication
 
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...IAEME Publication
 
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...IAEME Publication
 
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENTA MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENTIAEME Publication
 

Mais de IAEME Publication (20)

IAEME_Publication_Call_for_Paper_September_2022.pdf
IAEME_Publication_Call_for_Paper_September_2022.pdfIAEME_Publication_Call_for_Paper_September_2022.pdf
IAEME_Publication_Call_for_Paper_September_2022.pdf
 
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
 
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURSA STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
 
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURSBROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
 
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONSDETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
 
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONSANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
 
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINOVOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
 
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
 
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMYVISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
 
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
 
GANDHI ON NON-VIOLENT POLICE
GANDHI ON NON-VIOLENT POLICEGANDHI ON NON-VIOLENT POLICE
GANDHI ON NON-VIOLENT POLICE
 
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
 
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
 
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
 
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
 
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
 
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
 
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
 
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
 
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENTA MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
 

Último

A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI AgeCprime
 
Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Hiroshi SHIBATA
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...AliaaTarek5
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationKnoldus Inc.
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Scott Andery
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesKari Kakkonen
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better StrongerModern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better Strongerpanagenda
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesThousandEyes
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersNicole Novielli
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersRaghuram Pandurangan
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
Decarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityDecarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityIES VE
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc
 

Último (20)

A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI Age
 
Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog Presentation
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examples
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better StrongerModern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software Developers
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information Developers
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
Decarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityDecarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a reality
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
 

50120130406005

  • 1. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING & ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME TECHNOLOGY (IJCET) ISSN 0976 – 6367(Print) ISSN 0976 – 6375(Online) Volume 4, Issue 6, November - December (2013), pp. 48-61 © IAEME: www.iaeme.com/ijcet.asp Journal Impact Factor (2013): 6.1302 (Calculated by GISI) www.jifactor.com IJCET ©IAEME OPTICAL CHARACTER RECOGNITION (OCR) SYSTEM FOR MULTIFONT ENGLISH TEXTS USING DCT & WAVELET TRANSFORM Dr. MUSTAFA DHIAA AL-HASSANI Computer Science/Mustansiriyah University, Baghdad, Iraq ABSTRACT Optical Character Recognition (OCR) is a type of computer software designed to translate images of handwritten or typewritten text (usually captured by a scanner or a camera) into machineeditable text by recognizing characters at high speeds one at a time. OCR began as a field of research in pattern recognition, artificial intelligence and machine vision. It is becoming more and more important in the modern world according to economic reasons and business requirements. It helps humans ease their jobs and solve more complex problems by eliminating the time-consuming spent by human operators to re-type the documents and reduce error-prone processes. The presence of any type of noise or a combination of them can severely degrade the performance of OCR system. Though, a number of preprocessing techniques are considered in the present work in order to improve the obtained accuracy of the recognized text. An OCR system for 3185 training samples and 13650 testing samples is presented for multi-font English texts. Experiments have shown that wavelet features produce better recognition rates 96% than DCT features 92%. An improvement overall recognition rates (about 3%) are obtained after classifying characters according to the proportion of Height to Width feature to produce 99% for wavelet and 95% for DCT. Keywords: DCT, Feature Extraction, Optical Character Recognition (OCR), Pattern Recognition, Segmentation, Wavelet Transform. 1. INTRODUCTION Since the start of the computing era, information has been represented digitally so that it can be processed by computers. Approximately, more than 200 million paper books are being published yearly. Paper books and documents were abundant and widely being published at that time; and hence, there was a need to convert them into digital format. OCR was invented to translate the traditional paper-based books into digital e-books(i.e., electronic files). It was estimated that over 2 48
  • 2. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME million e-books are available for download on the Internet. E-books require less storage space than paper books, they can also be replicated many times, shared online, and digitally processed easily, mainly searched, translated, edited, and annotated. OCR systems are not that perfect as they are erroneous and exhibit spelling errors in the recognized output text, especially when the images being scanned are of poor printing quality [1, 2]. OCR is one of the most fascinating and challenging areas of pattern recognition with various practical applications: Automated postal address reading, ZIP code reading, checks, payment slips, income tax forms, business forms, automatic car plate-number recognition, and it can be used as an aid for visually handicapped people when combined with speech synthesizer [1, 3, 4]. Automatic character recognition is a subfield of pattern recognition and can either to be online or off-line. On-line recognition refers to those systems where the data to be recognized is input through a tablet digitizer, which acquires in real-time the position of the pen tip as the user writes. In contrast, off-line systems input the data from a document through an acquisition device, such as a scanner or a camera. Off-line character recognition is moreover divided into two categories: machine printed and handwritten [5]. The printed texts include all the printed materials such as: books, newspapers, magazines, and documents which are the outputs of typewriters, printers or plotters. OCR systems of machineprinted documents can be classified into [5, 6]: • Mono-font OCR systems that deals with documents written with one specific font, • Multi-font OCR systems that handles a subset of the existing fonts (recognition of more than one font), • Omni-font OCR systems that allows the recognition of characters in any font. Today many types of OCR software available in the markets like: Desktop OCR, Server OCR, Web OCR etc. Accuracy rate of any OCR tool varies from 71% to 98% [7]. 2. PROBLEM DEFINITION In modern society, we rely heavily on computers to process huge volumes of data. Related to this and for economic reasons or business requirements, there is a great demand for quickly converting the printed information in a document into an edited text in the computer. Often these data exist on paper and they have to be typed into the computer by human operators. Such timeconsuming and error-prone processes have been lightened by the invention of OCR systems [1]. Unfortunately, OCR systems are still erroneous and inaccurate, especially when the source document is of low printing quality [2]. Therefore, accuracy of these systems can be dependent on text preprocessing and segmentation algorithms. Sometimes it is difficult to retrieve text from the image because of different size, style, orientation, complex background of image … etc which produce misspellings in the recognized text [2, 7]. OCR technology allows machine to recognize the text automatically in the same way as the combination of eye and mind of human body. In development of computerized OCR system, few problems can occur [7]: • There is very little visible difference between some letters and digits for computers to understand. For example it might be difficult for the computer to differentiate between digit '0' and letter 'o' / 'O', • It might be very difficult to extract text, which is embedded in very dark background or printed on other words or graphics. 49
  • 3. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME 3. AIM OF THE WORK This paper aims to build an OCR Multi-font system which converts the printed English texts in a paper document (optical patterns exist in a digital image) into an edited text (its corresponding alphanumeric form or other symbols) in the computer in order to: • eliminate the time-consuming spent by human operators to re-type the documents (the huge volumes of data), • reduce the possible errors occurred in the typing process, • save money (by cancelling the need of human typists), • preserve space needed for paper books. 4. THE PROPOSED OCR SYSTEM The following block diagram, shown in Figure (1), illustrates the proposed OCR system model. Figure (1): Block-Diagram of the Proposed OCR System Model The input scanned image text is passed through a sequence of preprocessing steps (noise removal, foreground/ background separation, normalization, and binarization) prior to characters segmentation phase. Then, feature extraction methods (Discrete Cosine Transform (DCT) followed by Zigzag process, or Wavelet Transform (WT)) are applied to the segmented characters. The obtained feature-set is either stored in database as templates or references during the training phase when building database (DB) of characters features-set or is compared directly during the testing phase to those DB references in a pattern matching stage. Finally, decision rule is applied to produce recognition results beside the best matched characters. 50
  • 4. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME 4.1 The Input Scanned Image The input documents used for training and testing are scanned and digitized by a page scanner at 300 dpi resolution connected to a computer system and saved in BMP format of 256 graylevels. A number of scanned images that are used as inputs to our OCR system model are shown in Figure (2) below: Figure (2): 3–Samples of Scanned Images 4.2 Document Image Preprocessing Digital Images are generally corrupted by noise during the process of acquisition and transmission. This noise degrades the quality of digital image which produces several tiny dots scattered due to uneven gray-scale intensities which causes poor recognition rates. Consequently the performance of any system manipulating these images is also decreased. Therefore, removal of noise in document images corrupted by Gaussian and Impulse noises before OCR is important to guarantee better accuracy of characters recognition [1, 8]. Thus, image enhancement techniques adopted in this system model in the following sequence are employed prior to segmentation in order to simplify the process of characters segmentation [9, 10]: a) Noise Removal: is applied to the scanned document images for two primary purposes: to eliminate the noise and to give an image a softer effect. The spatial convolution mask of Gaussian filter used for low-pass filtering is shown in Figure (3). Figure (3): Gauss core – weighted on distance (Gaussian filter) Gaussian filter smoothen the image to match the pixelsnearby in a way that no point in the image differ from its surroundings to agreater extent. Image smoothing is accomplished in the spatial domain toremove errors, incorrect data and simplify the acquisition process of patterns [8, 10, 11]. 51
  • 5. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME b) Foreground/ Background Separation:is the process of separating the foreground regions (the area of interest containing the printed-text) in the image from the background regions (the useless area outside the borders of printed-text). The background regions generally exhibit a very low gray-scale variance value, whereas the foreground regions have a very high variance. Hence, a method based on variance thresholding can be used to perform the separation. Firstly, the image is divided into blocks and thegray-scale variance is calculated for each block in the image. If the variance is less than the global threshold, then the block is assigned to be a background region; otherwise, it is assigned to be part of the foreground. The gray-level variance for a block of size W × W is defined as [9, 11]: 1 V (k ) = W 2 W −1 W −1 ∑ ∑ ( I (i, i=0 j) − M (k ) )2 …… (1) j=0 where V(k) is the variance for kth block, I(i, j) is the gray-level value at pixel(i, j), and M(k) is the mean gray-level value for the kth block. c) Normalization: is utilized to reduce the effect of non-uniform intensities and improving image quality by stretching its histogram. To be able to normalize an image the area which is to normalize within has tobe known. Thus it is necessary to find the highest and the lowest pixel value ofthe current image. Every pixel is then evenly spread out along the scale by thefollowing equation [10, 11]: N (i, j ) = I ( i , j ) − I min × M I max − I min …… (2) whereI(i, j) is the gray-level value at pixel (i, j), Imin is the smallest gray-level value found in the image, Imax is the largest gray-level value found in the image, Mrepresents the new maximum gray-level value of the scale (i.e., M=255), and N(i, j) represent the normalized gray-level value at pixel (i, j). d) Binarization:Is the process of turning a gray scale image into a binary image (only two levels of interest 0 and 1) in order to improve the contrast, and consequently facilitates the feature extraction process.It is impossible to find a working global threshold value that can be used efficiently on every image because of the variations among the scanned images.Therefore algorithms to find the optimal value, based on localized thresholds, must be applied separately on each image to get a functional binarization. The image is partitioned into smaller blocks and threshold values are then calculated for each of these blocks. This enables adaptations that are not possible with global calculations. Localized thresholds demand a lot more calculations but mostly compensate it with better results [9, 11].The local mean threshold for kthblock of size W × W is computed below: 1 W −1 W −1 Local Mean ( k ) = 2 ∑ ∑ Block (i , j ) W i =0 j =0 …… (3) whereBlock (i, j) is the gray-level value at pixel (i, j).If the pixel value is lower than the threshold then the pixel is assigned to be part of the Printed-text; otherwise, it is assigned to be part of Background. 52
  • 6. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME Figure (4) illustrates the effect of applying the document image preprocessing techniques to a scanned document sample. Figure (4): The Effects of the proposed preprocessing steps on a scanned document image sample From Figure (4.c), it is noticeable that background regions are specified by black color, which is only to indicate the effect of this process despite the fact that in practical application the background is represented by white color. 4.3 Characters Segmentation Segmentation is an important phase and the accuracy of any OCR heavily depends on it, where incorrect segmentation leads to reduction in recognition accuracy. Segmentation is the process thatdivides the whole document into smaller components, which include [6, 12]: • • Line, Word, and Character segmentation The procedure adopted in this work for analyzing imagesto detect characters, as shown in Figure (5), is listed in the following sequence: Step1: Perform "Row Scan" (Line segmentation) to find the number of lines and boundaries of each line in any input document image within which the detection can proceed. 53
  • 7. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME Algorithm1 a) b) c) d) e) f) g) h) i) j) Get image data from the processed image. Initialize lines boundaries (T: Top, B: Bottom) to -1. Perform row scan (from 1st to last row) for pixels value= 0 (i.e, black). Set the number of lines to 0. If black pixel is detected then register T as the top of the current line and move the pointer to the next row & 1st column. Otherwise, continue to the next pixel (from left to right). If black pixel is found and T <> -1 then update B as the bottom of the current line and move the pointer to the next row & 1st column. If no black pixel is found in the row and T <> -1 then increment the number of lines by 1. Start below the bottom of the last line found and repeat steps e) – g) to detect subsequent lines (Stop when the bottom of image is reached). Process dot "." problem spacing found in (i and j) characters by merging lines below certain space threshold according to the font type & size and decrement the number of lines by 1 (if such case occurred). Print out the number of lines detected and draw line boundaries on the processed image for each detected lines (for: T–1 and B+1). Step2: Perform "Column Scan" (orthogonally from 1st to last column) only for detected lines. Thus, detecting characters in an image does not necessarily involve scanning the whole image all over again. Algorithm2 a) b) c) d) e) f) g) h) Initialize characters boundaries (L: Left, R: Right) to -1. Perform column scan (from 1st to last column) for pixels value= 0 (i.e, black). Set the number of characters to 0. If black pixel is detected then register L as the left of the current character and move the pointer to the next column & 1st row. Otherwise, continue to the next pixel (from top to bottom). If black pixel is found and L <> -1 then update R as the right of the current character and move the pointer to the next column & 1st row. If no black pixel is found in the column and L <> -1 then increment the number of characters by 1. Scan up to the right of the character found and repeat steps d) – f) to detect subsequent characters (Stop when the right-end of the last line is reached). Print out the number of characters detected and draw line boundaries on the image for each detected characters (for:L–1 andR+1). Step3: Perform "Row Scan"once more on the results obtained from the previous step in order to detect the actual character (top & bottom) boundaries. Step4: Bitmap imagesare created on the hard-disk for each segmented character relative to its boundaries; and its header information is generated from the original scanned image header besides updating dimensions. 54
  • 8. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME Step 2 Character bounds Left & Right Step 1 Line Top 1 Line Bottom 2 Step 3 Character Top Character Bottom 3 4 Figure (5): Lines and Characters boundary detection for a scanned image sample From the above figure, it is obvious that the detected lines-bound (top & bottom) stated by the "red-color" might not necessarily be the actual bounds for the characters in the same line because the characters have different heights. Hence, a confirmation of top and bottom boundaries for each character is needed as stated by the "green-color". The "blue-color" illustrates the detected characters-bound (left & right) for different characters widths. 4.4 Database Construction In general, any OCR system depends on training samples as input data. In this work, database of (91) samples were collected from the data sets shown in Table (1) below for different font types ("Arial", "Calibri", "Courier New", "Lucida Sans" and "Times New Roman") and for different font sizes (8, 10, 12, 14, 16, 18, 20) to produce (3185) samples. No. 1 2 3 4 Table (1): Data-Set Samples No. of Data Sets Samples Samples Digits Capital English Letters Small English Letters Some common ASCII Symbols 10 0 1 2 … 9 26 'A' 'B' 'C' … 'Z' 26 'a' 'b' 'c' … 'z' 29 .,;'":[]-+ */=(){} <> ! @ # $ % ^ & ? _ 4.5 Feature Extraction Feature extraction is part of the data reduction process by forming a new “smaller” set of features from the original feature set of the patterns. This can be done by extracting some numerical 55
  • 9. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME measurements from raw input patterns. Image features are of major importance in the isolation of regions of common property within an image[9, 10, 13]. In this work, two-sets of features were extracted from the segmented characters either by the use of Discrete Cosine Transform (DCT) or the spectral properties of Wavelet transform. a) Discrete Cosine Transform (DCT) DCT has become a standard method for many image processing& video compression algorithms. The two-dimensional DCT can be computed using the one-dimensional DCT horizontally (row-wise) and then vertically (column-wise) across the image because DCT is a function that separates the image into frequencies with large variance. The two-dimensional Forward DCT (2D FDCT) coefficients F(u, v)of M x N block of image pixels f(x , y)are formulated as [11, 13, 14, 15]: …… (4) …… (5) where C(k) is the normalization constant, u = 0,1,. . . , N – 1 and v = 0,1,. . . , M – 1. The DCT coefficients (Di,j and i, j = 0,1, … 7) of the corresponding image block of size 8×8, as example, are then ordered in a particular irregular sequence as shown in Figure (6). This irregular ordering of the coefficients is called Zig-zag ordering. Figure (6): Zig-zag ordering of DCT Coefficients The above sequence is broken into runs of nonzero (the early coefficients that contains the important "low-frequency" image information) and zero coefficients (the later coefficients in a block that contains the less-important "high-frequency" image information) [14, 15].Therefore, the final DCT feature-set considered in this work is generated only from the number of significant (nonzero) coefficients denoted by N in Table (2) that starts from D0,0. b) Wavelet Transform The wavelet transform has been found very useful for the time-scale representation and has been widely used in signal processing and computer vision. The Wavelet transform is a multiresolution technique that cut up data into different frequency components, and then analyzes each component with a resolution matched to its scale. The forward and inverse continuous wavelet 56
  • 10. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME transform of x(t)"the signal to be analyzed" with respect to the basis function or wavelet ψ j , k (t) at scale j (j>0) and time delay k is written as follows [16, 17, 18]: Forward CWT: W( j,k ) = Inverse CWT : where ∫ x(t) ψ (t) dt …… (6) x(t) = ∫ ∫ W(j,k) ψ j, k (t) dk dj …… (7) j,k k j ψ j,k (t) = 1 ψ( t − k ) j j and ψ(t) is the mother wavelet …… (8) This multiresolution can also be obtained using filter banks, resulting in the Discrete Wavelet Transform (DWT) that are well suited to the digital computer because there are no derivatives or integrals, just multiplication and addition operations which correspond to the mathematical convolution operation. The procedure starts with passing the signal (sequence) through a half-band digital low-pass & high-pass filters. The DWT is computed by successive low pass and high pass filtering of the discrete time-domain signal X[n], as shown in Figure (7), and each result is downsampled by two (↓2) where the low-pass filter is denoted by G0 while the high-pass filter is denoted by H0. At each level, the high-pass filter produces detail information d[n] while the low-pass filter associated with scaling function produces coarse approximations a[n].The DWT of the original signal is then obtained by concatenating all the coefficients, a[n] and d[n], starting from the last level of decomposition [10, 17, 18]. Figure (7): Three Levels DWT Decomposition Tree After the conversion of input image from its lowest-level of pixel data in spatial domain I(x, y) into higher-level representation of wavelet coefficients W(x, y), a set of wavelet features (the energy of each band as stated in Eq.(9))can be extracted by recursively decomposing sub-images in the low frequency channels as shown in Figure (8).The number of wavelet features for the 1st level is 4, and each progressing in wavelet level will correspond an increasing in features length by 3. …… (9) 57
  • 11. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME Figure (8): Three-Level Wavelet Multiresolution Decomposition Image Wavelet transform breaks an image down into four sub-sampled images. The results consist of one image that has been high pass filtered in both horizontal and vertical directions (HH), one that has been high pass filtered in the vertical and low pass filtered in the horizontal (LH), one that has been low passed in the vertical and high passed in the horizontal (HL), and one that has been low pass filtered in both directions(LL). Numerous filters can be used to implement the wavelet transform. Daubechies (D4) wavelet is one of the most commonly used due to its efficiency. The Daubechies basis vectors are [10, 16, 18]: …… (10) …… (11) 4.6 Pattern Matching The resulting test template, which is an N-dimensional feature vector, is compared against the stored reference templates to find the closest match. The process is to find which unknown class matches a predefined class or classes. For the OCR task, the unknown character is compared to all references in the database. This comparison can be done through Euclidean (E.D.) distance measure, shown below [9, 10]: …… (12) N E .D . = ∑ i =1 (a i − b i )2 where A and B are two vectors, such that A=[a1a2… aN]andB=[b1 b2… bN]. In our approach the minimum distance classifier is used to measure the difference between the two patterns (feature vectors). This classifier assigns the unknown pattern to the nearest predefined pattern. The bigger distance between the two vectors, is the greater difference [9, 10]. 5. EXPERIMENTAL RESULTS OCR accuracy is definedas the ratio of correct recognized characters to the total number of characters (samples) tested,as shown by Eq. (13): 58
  • 12. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME OCR Accuracy = No. of Correctly Re cognized Characters × 100 % Total No. of Characters Tested ..… (13) A number of experiments and test conditions were accomplished on (13650) samples to measure the performance of the proposed OCR system on various types of document images of different dimensions, font types and font sizes. As a result, the database size of training samples is computed as follows: No . of Training Samples = No . of Samples × No . of Fonts × No . of Font Sizes ..… (14) = 91× 5 × 7 = 3185 Samples A more appropriate comparison can be made if both DCT & Wavelet methods are measured under identical conditions. Based on the results shown in Table (2), one can deduce that all wavelet features produce better recognition rates 96% (from 92 – 98%) than DCT features 92% (from 87 – 95%). Different number of DCT coefficients (N) and wavelet decomposition levels values are examined according to the recognition rates. It is clearly indicated that two decomposition levels are most appropriate for wavelet feature vector construction, whereas 10-DCT coefficients is the suitable number of DCT features (N=10). Table (2): OCR Accuracy for different Testing images using DCT &Wavelet features Test Image File Font Type No. of Chars DCT Wavelet Transform Test1.bmp Test2.bmp Test3.bmp Arial Test4.bmp Test5.bmp Test6.bmp Test7.bmp Test8.bmp Calibri Test9.bmp Test10.bmp Test11.bmp Test12.bmp Courier Test13.bmp New Test14.bmp Test15.bmp Test16.bmp Test17.bmp Lucida Test18.bmp Sans Test19.bmp Test20.bmp Test21.bmp Test22.bmp Times Test23.bmp New Test24.bmp Roman Test25.bmp 525 581 791 371 462 480 633 415 692 510 395 704 620 557 454 664 325 653 341 747 720 373 475 525 637 N=5 77.52 79.00 81.29 75.74 71.65 76.46 79.78 78.31 79.05 77.06 75.70 81.82 79.35 79.35 75.11 80.72 72.92 78.56 68.62 79.65 76.11 71.05 73.68 79.05 78.02 N=10 91.62 93.46 90.90 94.88 87.88 92.29 93.84 95.66 90.75 92.75 92.41 95.60 92.58 93.36 91.63 93.67 94.46 92.04 89.15 91.30 90.28 87.13 89.05 93.14 90.74 N=15 85.90 87.44 85.59 83.83 78.35 82.29 85.47 84.10 83.38 84.31 82.28 86.36 83.23 84.02 81.28 86.30 82.15 82.54 77.71 82.33 79.44 80.43 79.79 82.86 81.48 Level1 71.43 74.87 80.28 64.15 69.05 75.63 77.09 68.43 75.29 79.02 67.34 73.86 75.97 74.33 68.06 75.75 68.92 76.11 69.21 78.71 77.64 72.12 68.63 74.10 78.96 Level2 94.29 96.73 94.82 97.30 93.51 95.00 95.10 96.63 96.68 97.25 96.46 97.30 96.61 98.20 96.92 95.63 92.62 94.49 95.60 94.38 96.81 98.93 95.37 96.95 97.02 Level3 80.95 87.78 85.59 80.32 83.77 83.75 80.73 83.37 85.55 84.90 81.27 84.52 83.71 89.23 82.60 84.19 79.38 82.24 78.01 83.53 85.00 86.33 80.63 85.52 86.81 Total 13650 77.64 92.05 83.16 74.25 96.01 83.89 59
  • 13. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME A good feature set has a major importance in the isolation of regions of common property within an image, and it should represent characteristic of a class that help distinguish it from other classes, as shown in Figure (9).The proportion of (‫ ) ܹ⁄ ܪ‬was taken in our consideration for all text components (characters, digits, and symbols) for two reasons: • it is invariant to different font sizes of the same font type, • to speed-up the recognition process because the matching processes are limited only to the text components in the same class. Consequently, an improvement overall recognition rates (about 3%) are obtained after classifying characters according to the addition of a new discriminating feature (the proportion of Height to Width) that produce 99% for wavelet and 95% for DCT. Figure (9): The Proportion of (۶⁄‫ )܅‬Feature for the character sample 'E' in different Font types & sizes 6. CONCLUSIONS A multi-fonts English texts OCR system for 3185 training samples and 13650 testing samples is presented that relies on DCT and wavelet features. Image enhancement techniques (noise removal, foreground/ background separation, normalization and binarization) are adopted in this work prior to segmentation in order to improve the recognition rates and simplify the process of characters segmentation. It is found that wavelet method is appropriate for feature vector construction where all the recognition rates (96%) outperform the DCT based recognition method (92%). To enhance the recognition rates further and speeds up the recognition process, text-components (characters, digits, and symbols) are classified according to the proportion of (‫ ) ܹ⁄ ܪ‬feature that produce 99% accuracy for wavelet based method because it help distinguish a class from other classes which is invariant to different font sizes of the same font type. REFERENCES [1] [2] Mohamed Cheriet, NawwafKharma, Cheng-Lin Liu, ChingY.Suen, "Character Recognition Systems: A Guide for Students and Practitioners", John Wiley & Sons, Inc., Canada, 2007. Youssef Bassil, Mohammad Alwani, "OCR Context-Sensitive Error Correction Based on Google Web 1T 5-Gram Data Set", American Journal of Scientific Research, ISSN 1450-223X, Issue. 50, February 2012. 60
  • 14. International Journal of Computer Engineering and Technology (IJCET), ISSN 0976-6367(Print), ISSN 0976 - 6375(Online), Volume 4, Issue 6, November - December (2013), © IAEME [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] R. JagadeeshKannan, R. Prabhakar, "An Improved Handwritten Tamil Character Recognition System using Octal Graph", Journal of Computer Science 4 (7): 509-516, ISSN 1549-3636, 2008. R. C. Tripathi, Vijay Kumar, "Character Recognition: A Neural Network Approach", National Conference on Advancement Technologies – Information Systems & Computer Networks (ISCON) published in IJCA®, 2012. Nafiz Arica, "An Off-Line Character Recognition System for Free Style Handwriting", M.Sc. thesis submitted to the department of computer engineering, the graduate school of natural and applied sciences of the Middle East Technical University, 1998. Ibrahim Abuhaiba, "Arabic Font Recognition Based on Templates", the International Arab Journal of Information Technology, Vol. 1, No. 0, July 2003. Chirag Patel, Atul Patel, Dharmendra Patel, "Optical Character Recognition by Open Source OCR Tool Tesseract: A Case Study", International Journal of Computer Applications (0975 – 8887) Vol. 55 – No.10, October 2012. C. Patvardhan, A. K. Verma, C. V. Lakshmi, "Denoising of Document Images using Discrete Curvelet Transform for OCR Applications", International Journal of Computer Applications (0975 – 8887), Vol. 55 – No.10, October 2012. Rafael C. Gonzalez, Richard E. Woods, “Digital Image Processing”, Second Edition, PrenticeHall, Inc., New Jersey, U.S.A., 2007. S. E. Umbaugh, “Computer Vision and Image Processing”, Prentice-Hall, Inc., U.S.A., 1998. TinkuAcharya, Ajoy K. Ray, "Image Processing: Principles and Applications", John Wiley & Sons, Inc., New Jersey, U.S.A., 2005. Ashu Kumar, Simpel Rani Jindal, "Segmentation of handwritten Gurmukhi text into lines", International Conference on Recent Advances and Future Trends in Information Technology (iRAFIT) published in IJCA®, 2012. William K. Pratt, "Digital Image Processing", Fourth Edition, John Wiley & Sons, Inc., New Jersey, U.S.A., 2007. Milan Sonka, Vaclav Hlavac, Roger Boyle, "Image Processing, Analysis and Machine Vision", Third International Student Edition, Thomson Corporation, U.S.A., 2008. David Salomon, "Data Compression: The Complete Reference", Fourth Edition, SpringerVerlag London Limited 2007. C. S. Burrus, R. A. Gopinath and H. Guo, “Introduction to Wavelets and Wavelet Transforms”, Prentice-Hall, Inc., U.S.A, 1998. M. Kociołek, A. Materka, M. Strzelecki P. Szczypiński, "Discrete Wavelet Transform – Derived Features for Digital Image Texture Analysis", Proc. of International Conference on Signals and Electronic Systems, pp. 163-168, Poland, 2001. V. Jeengar, S.N. Omkar, A. Singh, "A Review Comparison of Wavelet and Cosine Image Transforms", I.J.Image, Graphics and Signal Processing, DOI, 2012. M. M. Kodabagi, S. A. Angadi and Chetana. R. Shivanagi, “Character Recognition of Kannada Text in Scene Images using Neural Network”, International Journal of Graphics and Multimedia (IJGM), Volume 4, Issue 1, 2013, pp. 9 - 19, ISSN Print: 0976 – 6448, ISSN Online: 0976 –6456. Dr. Mustafa Dhiaa Al-Hassani and Dr. Abdulkareem A. Kadhim, “Design a Text-Prompt Speaker Recognition System using Lpc-Derived Features”, International Journal of Information Technology and Management Information Systems (IJITMIS), Volume 4, Issue 3, 2013, pp. 68 - 84, ISSN Print: 0976 – 6405, ISSN Online: 0976 – 6413. Dr. Mustafa Dhiaa Al-Hassani, Dr. Abdulkareem A. Kadhim and Dr. Venus W. Samawi, “Fingerprint Identification Technique Based on Wavelet-Bands Selection Features (WBSF)”, International Journal of Computer Engineering & Technology (IJCET), Volume 4, Issue 3, 2013, pp. 308 - 323, ISSN Print: 0976 – 6367, ISSN Online: 0976 – 6375. 61