Kunte 2007
Kunte 2007
4. Gabor filters
Figure 1. Consonant-vowel composite char- A Gabor filter can be viewed as a sinusoidal plane of par-
acters of first consonant. ticular frequency and orientation, modulated by a Gaussian
envelope. It is defined as,
To summarize, Kannada alphabet has (i) 16 vowels, (ii) h(x,y) = s(x,y)g(x,y) (1)
34 consonants, each with 34 basic form and 16 consonant-
vowel composite letters giving rise to 578 (34 x 17) char- where s(x,y) is a complex sinusoid, known as carrier and
acters and (iii) 34 subscripts to a total of 628 (16+578+34) g(x,y) is a 2-D Gaussian shaped function, known as enve-
characters. The full set of Kannada alphabet is given in [9]. lope. The complex sinusoid is defined as,
203
197
Figure 2. Word script recognition (Kanada / English) in bi-script document.
204
198
Hence, the 16 composite characters The consonant-vowel composite characters of a conso-
of con- nant are all grouped together as a single class of character in
sonant can be identified by the set of 8 characters the first instance of classification. Later, a character belong-
and the 6 symbols/subscripts ing to such a group is further classified within its group by a
separate small network meant for classifying that consonant
.
group of characters, in the second stage of classification.
All the 34 consonants in Kannada have a set of
Hence, in this classification methodology, a character
16 consonant-vowel composite characters similar to that
is classified in either a single stage or two stages. The
shown in Figure 1. For each consonant group of charac-
consonant-vowel composite characters require two stages
ters, the combinations remains the same, represented by
of classification. All other main characters (i.e vowels and
the 8 individual classes of characters and 6 individual sym-
symbols) and subscripts are classified in single stage only
bols/subscripts, totally 312 [(34 x 8 + 6=278) + 34 sub-
(first stage).
scripts] individual classes for recognition. This results in
a large reduction in the number of classes to be identified 7.3.1. First-stage of classification At the first level of clas-
by the classifiers. sification, there are two separate neural networks, one for
main character classification and another for subscript clas-
7. Classification sification. The input character is fed to either the main clas-
sifier or the subscript classifier depending upon whether it
Multilayer Perceptrons are being used as classifiers in is a main character or subscript character, which is deter-
the proposed work because of their universal approximation mined at character segmentation stage [10]. The subscript
property and good generalization ability [13]. They are ad- classifier is a small neural network designed to classify 34
equate for classification problems involving a small number classes of subscripts. The main classifier is a neural network
of classes. However, in large-set character recognition prob- designed to classify 14 vowels, 4 symbols and 34 consonant
lems, the conventional one-stage neural network classifier groups to a total of 52 classes. The neural classifier used for
either does not converge or even if it converges takes a large classification of main characters in the first stage is referred
amount of time [14]. Separate neural classifiers are devel- to as the first stage main character classifier.
oped (i) for separation of Kannada and English words from 7.3.2. Second-stage of classification The second level of
bilingual document (ii) for recognition of English charac- main character recognition neural network is a group of 34
ters and (iii) for recognition of Kannada characters. small architecture neural classifiers, one for each consonant
group. They are referred as group classifiers. An appropri-
7.1. Classifier for script recognition ate group classifier is selected by the first stage main char-
acter classifier to further classify a consonant or its compos-
The network architecture includes 40 nodes in input ite character within its group. Since a consonant group is a
layer (corresponding to 40 Gabor features), 40 nodes in the set of 8 characters, the number of outputs of any group clas-
hidden layer and 2 nodes in output layer (corresponding to sifier is only 8.The number of individual character classes
two classes of scripts i.e. to identify Kannada and English that the first stage main character classifier should handle
scripts). This network is referred as Script classifier. thus reduces to just 52 classes, as against 274 main charac-
ters. Each group classifier should handle just 8 characters
7.2. Classifier for English character recog- resulting in a set of simple architecture neural networks suf-
nition ficient for handling the complete character set of Kannada
for recognition.
The network architecture includes 120/160/200 nodes
in input layer (corresponding to 120/160/200 wavelet fea- 8. Experimental results and discussions
tures), 60/80/100 nodes in the hidden layer and 52 nodes in
output layer (corresponding 52 classes of English charac- The bilingual document is scanned using flatbed scanner.
ters). Lines and words are separated after pre-processing. Gabor
features are extracted from the normalized word image and
7.3. Two-stage classifier for Kannada char- applied to Script classifier for word recognition.
acter recognition
8.1. Script identification
The proposed two-stage classification methodology to
classify the 312 individual classes of Kannada characters The obtain the data set for training, about 250 differ-
is as follows: ent samples of Kannada and English words are taken from
205
199
different Kannada news papers and magazines. The script classifier. The test samples corresponding to each conso-
identification classifier was tested with about 350 samples nant group are copied into separate test data files to test the
from Kannada and English words, which were different group classifiers independently. All the trained classifiers
from training set. Kannada and English words were sep- of the recognition system are tested independently by pre-
arated with an accuracy of about 99.3% and 98.6% re- senting the test patterns from their individual test data files.
spectively. Kannada words which are mostly circular have The first stage main character classifier could classify
showed nearly equal response to all the Gabor filter orien- the composite characters grouped together for a consonant
tations. Where as English words which have more vertical group as same class of character and as well differentiate
and horizontal strokes have showed more response to 00 and between the samples of different consonant groups, vowels
900 filter orientations. This discriminating feature is used to and symbols with an effective maximum rate of about 92%.
classify the two scripts namely; Kannada and English. The recognition rate of subscript classifier is about 94%.
The average error rate of 34 group classifiers is less than
8.2. Kannada/English character recogni- 5% when tested independently from their test data files.
tion Since, the correct recognition of any complete character
in Kannada involves correct identification of the group, and
The wavelet features of the character image are extracted then the identification of the character within the group, the
by applying Discrete Wavelet Transform (up to one level overall recognition rate at character level of the system is
of decomposition using Daubechies wavelet) to the pre- found to be about 91% considering first stage main charac-
processed, normalized and resampled contour of the char- ter classifier, subscript classifier and group classifiers.
acter. For each class of character 3 sets of wavelet features A comparison of recognition rates for different num-
are extracted corresponding to 60, 80 and 100 number of ber of contour points against different rejection threshold
points in the resampled character contour, for finding the is plotted in Figure 4(a) and Figure 4(b) for first stage main
optimal number of points required to represent the charac- character classifier and subscript classifier respectively. It
ter. can be seen that improvement in recognition rate for first
stage main character classifier is very less for 100 contour
8.2.1. Kannada character recognition The data set for points. Similarly there is less improvement in recognition
training the neural network used for recognizing Kannada rate for 80 and 100 contour points for subscript classifier.
characters is formed from, about 30 different samples of Hence, the optimal number of contour points for represent-
each class of Kannada character (with different fonts and ing the Kannada characters is chosen as 60 and 80 for the
font size varying from 14 to 18 points) totally about 9000 cases of subscripts and main (non-subscript) characters re-
characters. A test set is created with about 20 different sam- spectively.
ples (other than training samples) of each class of Kannada
character totally about 6000 characters scanned from differ- 8.2.2. English character recognition The data set for train-
ent Kannada newspapers and magazines. The architectures ing the neural network used for recognizing English char-
of the different classifiers used are as follows: acters is formed from, about 100 different samples of each
class of English character (with different fonts and font size
• The first stage main character classifier, used for varying from 10 to 16 points) totally about 5200 characters.
group identification consists of 120/160/200 nodes in A test set is created with about 200 different samples (other
the input layer (corresponding to 120/160/200 wavelet than training samples) of each class of English character
features), 60/80/100 nodes in hidden layer and 52 totally about 10400 characters. A maximum recognition
nodes in output layer (corresponding to 52 classes). rate of 98.2% is obtained. The recognition rate is higher
• The subscript classifier used for subscript identifica- than that of Kannada character as the number of classes and
tion consists of 120/160/200 nodes in the input layer, number of similar shaped characters is less in English script.
60/80/100 nodes in hidden layer and 34 nodes in out- Considering Script classifier, Kannada character classi-
put layer (corresponding to 34 subscripts). fier and English Classifier an overall recognition rate of
90.5% is obtained at character level.
• The 34 group classifiers used for identifying the char-
acter class within a particular consonant group consists
of 120/160/200 nodes in the input layer, 60/80/100
9. Conclusion
nodes in hidden layer and 8 nodes in output layer (cor-
responding to 8 characters in a consonant group). In this paper, a bilingual OCR system for complete set of
printed Kannada and English characters employing wavelet
Main character samples are all put together and stored in descriptors is presented. Gabor features are used to recog-
a separate test data file to test the first stage main character nize the script at word level. A two-stage multi-network
206
200
Figure 4. Recognition rate for (a) first stage main character classifier (b) subscript classifier.
scheme for classification of large-set of characters and re- recognition of non-uniform sized characters – An exper-
alization of this scheme for printed Kannada character is imentation with Kannada characters, Pattern Recognition
proposed. A procedure to reduce large set of Kannada char- Letters, Vol. 37, 1999, 1467–1475.
acters for classification is also developed. Even though the [5] B. Vijay Kumar and A.G. Ramakrishnan, Radial ba-
recognition rate of Script classifier and English character sis function and subspace approach for printed Kannada
classifier is high, at character level combining all the classi- text recognition, Proc. IEEE International Conference on
fiers a recognition rate of 90.5% is got due to lower recog- Acoustics, Speech, and Signal Processing, Vol. 5, 2004,
nition rate of Kannada character classifier. Further research 321–324.
is on the way to improve the recognition rate for Kannada [6] T. V. Ashwin and P. S. Sastry, A font and size–independent
characters by adding structural features. The proposed sys- OCR system for printed Kannada documents using support
tem was implemented on PIV, 3.2 GHz computer system vector machines, Sadhana, Vol. 27, No. 1, 2002, 35–58.
using VC++ under Windows environment.
[7] Y. Zhu, T. Tan and Y. Wang, Font Recognition Based on
Global Texture Analysis, IEEE Trans. PAMI, Vol. 23, No.
Acknowledgment 10, 2001, 1192–1200.
[8] I. Daubechies, Ten Lectures on Wavelets, CBMS-
This work was supported in part by research grants from Conference Lecture notes, 71, SIAM Philadelphia, 1992.
University Grants Commission (UGC), New Delhi, under
Major Research Project scheme, F. No. 32 - 113/2006(SR). [9] R. Sanjeev Kunte and Sudhaker Samuel R.D, Wavelet De-
scriptors for Recognition of Basic Symbols in Printed Kan-
nada Text, International Journal of Wavelets, Multiresolu-
References tion and Information Processing, Vol. 5, No. 2, 351–367.
[10] R. Sanjeev Kunte and Sudhaker Samuel R.D, A Two–stage
[1] B.B. Chaudhuri and U. Pal, An OCR system to read two Character Segmentation Technique for Printed Kannada
Indian language scripts: Bangla and Devanagari, Proc. Text, International Journal on Graphics, Vision and Image
of Fourth International Conference on Document Analysis Processing, Vol. 6, 2006, 1–8.
and Recognition, 1997, 1011–1015.
[11] Chungnan Lee and Bohom Wu, A Chinese character stroke
[2] Atul Negi, Chakravarthy Bhagavathi and Krishna B, An extraction algorithm based on contour information, Pattern
OCR system for Telugu, Proc. of Sixth International Con- Recognition, Vol. 31, No. 6, 1998, 651–663.
ference on Document Analysis and Recognition, 2001,
1110–1114. [12] P. Wunch and F. L. Andrew, Wavelet descriptors for Mul-
tiresolution recognition of handprinted characters, Pattern
[3] Arun K. Pujari, C. Dhanunjaya Naidu, M. Sreenivasa Rao, Recognition, Vol. 28, No. 8, 1995, 1237–1249.
B. C. Jinaga, An intelligent character recognizer for Telugu
[13] H. White, Artificial neural networks: Approximation and
scripts using multiresolution analysis and associative mem-
learning theory, Blackwell, Oxford, 1992.
ory, Image Vision Comput, Vol. 22, No. 14, 2004, 1221–
1227. [14] Hee-Heon Song and Seong-Whan Lee, A self-organizing
neural tree for large-set pattern classification, IEEE Trans.
[4] P. Nagabhushan and Radhika. M. Pai, Modified region de- on Neural networks, Vol. 9, No. 7, 1998, 369–380.
composition method and optimal depth decision tree in the
207
201