Javascript must be enabled to continue!
Comparative Study of Character Recognition for Handwritten Characters
View through CrossRef
Optical Character Recognition (OCR) represents an important technology in the context of computer vision, which facilitates the extraction of textual data with the help images. Although a lot of investigations have examined various OCR models, there is a notable absence of comparative studies evaluating different algorithms on a unified standardized dataset. This research targets to fill this void by assessing multiple OCR models across two distinct datasets: one consisting of 28×28-pixel images and the other comprising of 64×64-pixel images. The models evaluated include ten Convolutional Neural Networks (CNNs) characterized by diverse activation functions, architectural depths, and, dropout rates, in addition to Long Short-Term Memory (LSTM) networks, Support Vector Machines (SVM), Encoder-Decoder frameworks, and Random Forest classifiers.
Through our analysis, it was revealed that the CNN-based models demonstrate exceptional performance, with the leading 64×64 CNN model achieving an accuracy of 0.9882, while the highest-performing 28×28 CNN model reported an accuracy of 0.9763. The Encoder-Decoder model also showed formidable results, achieving an accuracy of 0.9781 on the 64×64 dataset and 0.9810 on the 28×28 dataset. SVM exhibited robust performance on the higher-resolution dataset, achieving an accuracy of 0.9777, but encountered significant challenges on the 28×28 dataset, where the accuracy of only 0.7252 was reported. The Random Forest classifier maintained a consistent accuracy of 0.9538 across both datasets. Conversely, LSTM models struggled to generalize effectively for OCR applications, with the best LSTM model achieving a mere 0.0346 accuracy on the 64×64 dataset and performing poorly on the 28×28 dataset as well.
Through the comprehensive study of CNN model, it was observed that, Model 3 attained the highest accuracy of 0.9907 across both datasets, accompanied by minimal validation loss (0.0377 for 28×28 and 0.0577 for 64×64). Other CNN models exhibited varying levels of performance, with deeper architectures generally surpassing their shallower counterparts. Our methodology encompassed preprocessing the datasets, partitioning them into training and testing sets, and training each model with suitable hyperparameters. The findings underscore that CNN-based architectures are the most effective for OCR tasks, particularly at elevated resolutions.
These results offer significant insights into the efficacy of various OCR models.
Title: Comparative Study of Character Recognition for Handwritten Characters
Description:
Optical Character Recognition (OCR) represents an important technology in the context of computer vision, which facilitates the extraction of textual data with the help images.
Although a lot of investigations have examined various OCR models, there is a notable absence of comparative studies evaluating different algorithms on a unified standardized dataset.
This research targets to fill this void by assessing multiple OCR models across two distinct datasets: one consisting of 28×28-pixel images and the other comprising of 64×64-pixel images.
The models evaluated include ten Convolutional Neural Networks (CNNs) characterized by diverse activation functions, architectural depths, and, dropout rates, in addition to Long Short-Term Memory (LSTM) networks, Support Vector Machines (SVM), Encoder-Decoder frameworks, and Random Forest classifiers.
Through our analysis, it was revealed that the CNN-based models demonstrate exceptional performance, with the leading 64×64 CNN model achieving an accuracy of 0.
9882, while the highest-performing 28×28 CNN model reported an accuracy of 0.
9763.
The Encoder-Decoder model also showed formidable results, achieving an accuracy of 0.
9781 on the 64×64 dataset and 0.
9810 on the 28×28 dataset.
SVM exhibited robust performance on the higher-resolution dataset, achieving an accuracy of 0.
9777, but encountered significant challenges on the 28×28 dataset, where the accuracy of only 0.
7252 was reported.
The Random Forest classifier maintained a consistent accuracy of 0.
9538 across both datasets.
Conversely, LSTM models struggled to generalize effectively for OCR applications, with the best LSTM model achieving a mere 0.
0346 accuracy on the 64×64 dataset and performing poorly on the 28×28 dataset as well.
Through the comprehensive study of CNN model, it was observed that, Model 3 attained the highest accuracy of 0.
9907 across both datasets, accompanied by minimal validation loss (0.
0377 for 28×28 and 0.
0577 for 64×64).
Other CNN models exhibited varying levels of performance, with deeper architectures generally surpassing their shallower counterparts.
Our methodology encompassed preprocessing the datasets, partitioning them into training and testing sets, and training each model with suitable hyperparameters.
The findings underscore that CNN-based architectures are the most effective for OCR tasks, particularly at elevated resolutions.
These results offer significant insights into the efficacy of various OCR models.
Related Results
Primerjalna književnost na prelomu tisočletja
Primerjalna književnost na prelomu tisočletja
In a comprehensive and at times critical manner, this volume seeks to shed light on the development of events in Western (i.e., European and North American) comparative literature ...
ON-LINE HANDWRITTEN ARABIC CHARACTER RECOGNITION BASED ON GENETIC ALGORITHM
ON-LINE HANDWRITTEN ARABIC CHARACTER RECOGNITION BASED ON GENETIC ALGORITHM
On-line Arabic handwritten character recognition is one of the most challenging problems in pattern recognition field. By now, printed Arabic character recognition and on-line Arab...
Filtering Approaches and Mish Activation Function Applied on Handwritten Chinese Character Recognition
Filtering Approaches and Mish Activation Function Applied on Handwritten Chinese Character Recognition
Handwritten Chinese Characters (HCC) have recently received much attention as a global means of exchanging information and knowledge. The start of the information age has increased...
Invarianceness for Character Recognition Using Geo-Discretization Features
Invarianceness for Character Recognition Using Geo-Discretization Features
<span style="font-size: 10pt; font-family: 'Times New Roman','serif'; mso-bidi-font-size: 11.0pt; mso-fareast-font-family: 宋体; mso-font-kerning: 1.0pt; mso-ansi-language: EN-US;...
Deformed character recognition using convolutional neural networks
Deformed character recognition using convolutional neural networks
Realization of high accuracies towards south Indian character recognition is one the truly interesting research challenge. In this paper, our investigation is focused on recognitio...
OPTIMIZING CNN HYPERPARAMETERS FOR ENHANCED HANDWRITTEN DIGIT RECOGNITION ON CUSTOM DATASET: A SYSTEMATIC STUDY
OPTIMIZING CNN HYPERPARAMETERS FOR ENHANCED HANDWRITTEN DIGIT RECOGNITION ON CUSTOM DATASET: A SYSTEMATIC STUDY
Handwritten Digit Recognition is still an essential issue in artificial intelligence and pattern recognition. Convolutional Neural Networks (CNNs) have shown outstanding accuracy o...
Hindi-Handwritten-Character-Recognition using Deep learning
Hindi-Handwritten-Character-Recognition using Deep learning
Abstract: Recognizing handwritten Hindi characters poses a significant challenge in the realms of machine learning and computer vision, particularly in the context of India's accel...
Unveiling Cross-Linguistic Mastery: Advancing Multilingual Handwritten Numeral Recognition with Attention-driven Transfer Learning
Unveiling Cross-Linguistic Mastery: Advancing Multilingual Handwritten Numeral Recognition with Attention-driven Transfer Learning
Abstract
In the realm of data analysis and document processing, the recognition of handwritten numerals stands as a pivotal advancement. This contribution has steered trans...

