Character recognition technique, associates a symbolic identity with the image of the character, is an important area in pattern recognition and image processing. The principal idea here is to convert raw images (scanned from document, typed, pictured etcetera) into editable text like html, doc, txt or other formats. There is a very limited number of Bangla Character recognition system, if available they can’t recognize the whole alphabet set. Motivated by this, this paper demonstrates a MATLAB based Character Recognition system from printed Bangla writings. It can also compare the characters of one image file to another one. Processing steps here involved binarization, noise removal and segmentation in various levels, features extraction and recognition.
Image to Text Converter PPT. PPT contains step by step algorithms/methods to which we can convert images in to text , specially contains algorithms for images which contains human handwritting, can convert writting in to text, img to text.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Fuzzy Region Merging Using Fuzzy Similarity Measurement on Image Segmentation IJECEIAES
Some image’s regions have unbalance information, such as blurred contour, shade, and uneven brightness. Those regions are called as ambiguous regions. Ambiguous region cause problem during region merging process in interactive image segmentation because that region has double information, both as object and background. We proposed a new region merging strategy using fuzzy similarity measurement for image segmentation. The proposed method has four steps; the first step is initial segmentation using mean-shift algorithm. The second step is giving markers manually to indicate the object and background region. The third step is determining the fuzzy region or ambiguous region in the images. The last step is fuzzy region merging using fuzzy similarity measurement. The experimental results demonstrated that the proposed method is able to segment natural images and dental panoramic images successfully with the average value of misclassification error (ME) 1.96% and 5.47%, respectively.
OBJECT SEGMENTATION USING MULTISCALE MORPHOLOGICAL OPERATIONSijcseit
Object segmentation plays an important role in human visual perception, medical image processing and content based image retrieval. It provides information for recognition and interpretation. This paper uses mathematical morphology for image segmentation. Object segmentation is difficult because one usually does not know a priori what type of object exists in an image, how many different shapes are there and what regions the image has. To carryout discrimination and segmentation several innovative segmentation methods, based on morphology are proposed. The present study proposes segmentation method based on multiscale morphological reconstructions. Various sizes of structuring elements have been used to segment simple and complex shapes. It enhances local boundaries that may lead to improve segmentation accuracy.The method is tested on various datasets and results shows that it can be used for both interactive and automatic segmentation.
This document discusses a system for extracting text from images. It begins with an introduction describing the need for such a system. It then covers related work on text detection techniques. The proposed method involves converting images to grayscale, binarization, connected component analysis, horizontal/vertical projections, reconstruction and using OCR for recognition. Applications discussed include wearable devices, video coding, image indexing and license plate recognition. While the system is robust, OCR recognition of noisy extracted text remains a challenge.
An effective approach to offline arabic handwriting recognitionijaia
Segmentation is the most challenging part of the Arabic handwriting recognition, due to the unique
characteristics of Arabic writing that allows the same shape to denote different characters. In this paper,
an off-line Arabic handwriting recognition system is proposed. The processing details are presented in
three main stages. Firstly, the image is skeletonized to one pixel thin. Secondly, transfer each diagonally
connected foreground pixel to the closest horizontal or vertical line. Finally, these orthogonal lines are
coded as vectors of unique integer numbers; each vector represents one letter of the word. In order to
evaluate the proposed techniques, the system has been tested on the IFN/ENIT database, and the
experimental results show that our method is superior to those methods currently available.
Zernike moment of invariants for effective image retrieval using gaussian fil...IAEME Publication
This document summarizes a research paper that proposes using Zernike moments and steerable Gaussian filters for effective image retrieval based on color, texture, and shape features. It describes extracting dominant colors from images using dynamic color quantization and clustering. Texture features are represented using steerable filter decomposition. Shape features are described using pseudo-Zernike moments. The proposed method combines these color, texture, and shape features to generate a robust feature set for image retrieval that provides better retrieval accuracy than other methods.
An improved double coding local binary pattern algorithm for face recognitioneSAT Journals
Abstract A human face conveys a lot of information about the identity and emotional state of the person. So now a day’s face recognition has become an interesting and challenging problem. Face recognition plays a vital role in many applications such as authenticating a person, system security, verification and identification for law enforcement and personal identification among others. So our research work mainly consists of three parts, namely face representation, feature extraction and classification. The first part, Face representation represents how to model a face and check which algorithms can be used for detection and recognition purpose. In the second phase i.e. feature extraction phase we compute the unique features of the face image. In the classification phase the computed DLBP face image is compared with the images from the database. In our research work, we use Double Coding Local Binary Patterns to evaluate face recognition which concentrate over both the shape and texture information to represent face images for person independent face recognition. The face area is firstly cut into small regions from which Local Binary Patterns (LBP), then we compute histograms to generate LBP image then we compute single oriented mean image from which we again compute histogram values small regions and at last concatenated into a single feature vectors and generate D-LBP image. This feature are used for the representation of the face and to measure similarities between images. Keywords: local binary pattern (LBP), double coding local binary pattern (D-LBP), features extraction, classification, pattern recognition, histogram, feature vector.
Text-Image Separation in Document Images Using Boundary/Perimeter DetectionIDES Editor
Document analysis plays an important role in office
automation, especially in intelligent signal processing. The
proposed system consists of two modules: block segmentation
and block identification. In this approach, first a document is
segmented into several non-overlapping blocks by utilizing a
novel recursive segmentation technique, and then extracts
the features embedded in each segmented block are extracted.
Two kinds of features, connected components and image
boundary/perimeter features are extracted. The features are
verified to be effective in characterizing document blocks.
Last, the identification module to determine the identity of
the considered block is developed. Wide verities of documents
are used to verify the feasibility of this approach.
Content-based Image Retrieval Using The knowledge of Color, Texture in Binary...Zahra Mansoori
This document presents a new approach for content-based image retrieval that combines color, texture, and a binary tree structure to describe images and their features. Color histograms in HSV color space and wavelet texture features are extracted as low-level features. A binary tree partitions each image into regions based on color and represents higher-level spatial relationships. The performance of the proposed system is evaluated on a subset of the COREL image database and compared to the SIMPLIcity image retrieval system. Experimental results show the proposed system has better retrieval performance than SIMPLIcity in some categories and comparable performance in others.
This document summarizes an evaluation of texture feature extraction methods for content-based image retrieval, including co-occurrence matrices, Tamura features, and Gabor filters. The evaluation tested these methods on a Corel image collection using Manhattan distance as the similarity measure. Co-occurrence matrices performed best with homogeneity as the feature, while Gabor wavelets showed better performance for homogeneous textures of fixed sizes. Tamura features performed poorly with directionality. Overall, co-occurrence matrices provided the best results for general texture retrieval.
EFFECTIVE SEARCH OF COLOR-SPATIAL IMAGE USING SEMANTIC INDEXINGIJCSEA Journal
Most of the data stored in libraries are in digital form will contain either pictures or video, which is tough to search or browse. Methods which are automatic for searching picture collections made large use of color histograms, because they are very strong to wide changes in viewpoint, and can be calculated trivially. However, color histograms unable to present spatial data, and therefore tend to give lesser results. By using combination of color information with spatial layout we have developed several methods, while retrieving the advantages of histograms. A method computes a given color as a function of the distance between two pixels, which we call a color correlogram. We propose a color-based image descriptor that can be used for image indexing based on high-level semantic concepts. The descriptor is
based on Kobayashi’s Color Image Scale, which is a system that includes 130 basic colors combined in 1180 three-color combinations. The words are represented in a two dimensional semantic space into groups based on perceived similarity. The modified approach for statistical analysis of pictures involves transformations of ordinary RGB histograms. Then a semantic image descriptor is derived, containing semantic data about both color combinations and single colors in the image.
Text-Image Separation in Document Images using Boundary/Perimeter DetectionIDES Editor
Document analysis plays an important role in office
automation, especially in intelligent signal processing. The
proposed system consists of two modules: block segmentation
and block identification. In this approach, first a document is
segmented into several non-overlapping blocks by utilizing a
novel recursive segmentation technique, and then extracts
the features embedded in each segmented block are extracted.
Two kinds of features, connected components and image
boundary/perimeter features are extracted. Document with
text inside image pose limitations in earlier reported literature.
This is taken care of by applying additional pass of the Run
Length Smearing on the extracted image that contains text.
Proposed scheme is independent of type and language of the
document.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
A Novel Document Image Binarization For Optical Character RecognitionEditor IJCATR
This paper presents a technique for document image binarization that segments the foreground text accurately from poorly
degraded document images. The proposed technique is based on the
Segmentation of text from poorly degraded document images and
it is a very demanding job due to the high variation between the background and the foreground of the document. This paper pr
oposes
a novel document image binarization technique that segments t
he texts by using adaptive image contrast. It is a combination of the
local image contrast and the local image gradient that is efficient to overcome variations in text and background caused by d
ifferent
types degradation effects. In the proposed technique
, first an adaptive contrast map is constructed for a degraded input document
image. The contrast map is then binarized by global thresholding and pooled with Canny’s edge map detection to identify the t
ext
stroke edge pixels. By applying Segmentation the
text is further segmented by a local thresholding method that. The proposed method
is simple, strong, and requires minimum parameter tuning
User Interactive Color Transformation between ImagesIJMER
Abstract: In this paper we present a process called color
transfer which can borrow one image’s color
characteristics from another. Most current colorization
algorithms either require a significant user effort or have
large computational time. Here focus on orthogonal color
space i.e. lαβ color space without correlation between the
axes is given. Here we have implemented two global color
transfer algorithms in lαβ color space using simple color
statistical information such as mean, standard deviation
and covariance between the pixels of image. Our approach
is the extension of Reinhard's. Our local color transfer
algorithm uses simple color statistical analysis to recolor
the target image according to selected color range in
source image. Target image’s color influence mask is
prepared. It is a mask that specifies what parts of target
image will be affected according to selected color range.
After that target image is recolored in lαβ color space
according to prepared color influence map. In the lαβ
color space luminance and chrominance information is
separate so it allows making image recoloring optional.
The basic color transformation uses stored color statistics
of source and target image. All the algorithms are
implemented in JAVA object oriented language. The main
advantage of proposed method over the existing one is it
allows the user to recolor a part of the image in a simple &
intuitive way, preserving other color intact & achieving
natural look.
Index Terms: color transfer, local color statistics, color
characteristics, orthogonal color space, color influence
map.
CONTRAST ENHANCEMENT TECHNIQUES USING HISTOGRAM EQUALIZATION METHODS ON COLOR...IJCSEA Journal
Histogram equalization (HE) is a simple and widely used image contrast enhancement technique. The basic disadvantage of HE is it changes the brightness of the image. In order to overcome this drawback, various HE methods have been proposed. These methods preserves the brightness on the output image but, does not have a natural look. In order to overcome this problem the, present paper uses Multi-HE methods, which decompose the image into several sub images, and classical HE method is applied to each sub image. The algorithm is applied on various images and has been analysed using both objective and subjective assessment.
The document presents a novel method for character segmentation of vehicle license plates written in two rows. It discusses how the license plate image is first preprocessed through steps like grayscale conversion, binarization, and noise removal. Horizontal and vertical projections are then used to segment the image into lines and words. Character segmentation is done by analyzing the spacing between characters. Finally, zone segmentation is performed to divide each character into four zones to extract features for recognition. The method aims to simplify license plate image analysis for applications like traffic monitoring and surveillance.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
This document presents a hybrid approach for color image segmentation that integrates color edge information and seeded region growing. It uses color edge detection in CIE L*a*b color space to select initial seed regions and guide region growth. Seeded region growing is performed based on color similarity between pixels. The edge map and region map are fused to produce homogeneous regions with closed boundaries. Small regions are then merged. The approach is tested on images from the Berkeley segmentation dataset and produces reasonably good segmentation results by combining color and edge information.
This document summarizes a research paper that proposes using fuzzy logic and an artificial bee colony (ABC) optimization technique to enhance degraded color images. The approach first converts image pixels from RGB to HSV color space. It then fuzzifies the value and saturation components using Gaussian membership functions while preserving hue. The ABC optimization technique is used to optimize the membership function parameters to improve convergence time and image quality. Experimental results showed this approach yielded better results than one using ant colony optimization, converging faster while enhancing image clarity.
This document proposes a new method for segmenting outdoor images called Color Cluster Elimination (CCE) which utilizes color clustering and texture analysis. CCE performs color clustering in a multi-resolution pyramid to gradually eliminate larger color clusters, preventing them from dominating segmentation and allowing smaller clusters to emerge more clearly. It then examines regions for adjacent homochromatic objects with different textures, introducing Texture Sewn Response (TSR) to indicate texture strength across resolutions/directions. The method is evaluated on the BSDS500 dataset against other metrics, demonstrating satisfactory performance for outdoor scene segmentation.
Evaluation of Euclidean and Manhanttan Metrics In Content Based Image Retriev...IJERA Editor
This document evaluates the performance of the Euclidean and Manhattan distance metrics in a content-based image retrieval system. It finds that the Manhattan distance metric showed better precision than the Euclidean distance metric. The system uses color histograms and Gabor texture features to represent images. Color is represented in HSV color space and histograms of hue, saturation and value are used. Gabor filters are applied to capture texture at different scales and orientations. Distance between feature vectors is calculated using Euclidean and Manhattan distance formulas to find similar images from the database. The system was tested on a dataset of 1000 Corel images and Manhattan distance produced more relevant search results.
This document discusses the application of morphological image processing in forensics for fingerprint enhancement. It provides background on morphological operations like dilation, erosion, opening and closing. It explains how these operations can be used to enhance degraded fingerprints by thickening ridges, joining broken ridges, and separating overlapped ridges. The morphological image processing concepts are implemented in Java to experimentally enhance fingerprint images and reduce noise.
The document describes an algorithm for detecting text in camera-captured images. It begins with preprocessing steps like converting the color image to grayscale, applying edge detection and morphological operations like dilation and erosion. This gives initial bounding boxes containing candidate text regions. Further processing includes applying geometrical constraints to filter boxes, performing multiresolution analysis, connected component analysis and filtering by area to get the final text regions. Inversion and addition steps are used to handle text against different backgrounds.
A comparative study on content based image retrieval methodsIJLT EMAS
Content-based image retrieval (CBIR) is a method of
finding images from a huge image database according to persons’
interests. Content-based here means that the search involves
analysis the actual content present in the image. As database of
images is growing daybyday, researchers/scholars are searching
for better techniques for retrieval of images maintaining good
efficiency. This paper presents the visual features and various
ways for image retrieval from the huge image database.
Nowadays character recognition has gained lot of attention in the field of pattern recognition due to its application in various fields. It is one of the most successful applications of automatic pattern recognition. Research in OCR is popular for its application potential in banks, post offices, office automation etc. HCR is useful in cheque processing in banks; almost all kind of form processing systems, handwritten postal address resolution and many more. This paper presents a simple and efficient approach for the implementation of OCR and translation of scanned images of printed text into machine-encoded text. It makes use of different image analysis phases followed by image detection via pre-processing and post-processing. This paper also describes scanning the entire document (same as the segmentation in our case) and recognizing individual characters from image irrespective of their position, size and various font styles and it deals with recognition of the symbols from English language, which is internationally accepted.
SEGMENTATION OF CHARACTERS WITHOUT MODIFIERS FROM A PRINTED BANGLA TEXTcscpconf
This document discusses the segmentation of printed Bangla characters without modifiers for optical character recognition systems. It begins with an introduction to OCR systems and Bangla script. The main steps of an OCR system are then outlined, with a focus on the segmentation step. Line, word and character segmentation algorithms are described in detail along with figures to illustrate the steps. The goal is to properly segment individual characters for recognition.
This document describes an extended fuzzy c-means (EFCM) clustering algorithm for noisy image segmentation. The algorithm first preprocesses noisy pixels in an image by regenerating their values based on neighboring pixel intensities. It then applies the conventional fuzzy c-means clustering algorithm to segment the image. The EFCM approach is presented as being less sensitive to noise than other clustering algorithms and able to efficiently segment noisy images. The document provides background on image segmentation, fuzzy c-means clustering, types of image noise, and density-based clustering challenges. It also outlines the EFCM methodology and its computational advantages over other robust clustering methods for noisy image segmentation.
Feature Extraction and Feature Selection using Textual Analysisvivatechijri
After pre-processing the images in character recognition systems, the images are segmented based on
certain characteristics known as “features”. The feature space identified for character recognition is however
ranging across a huge dimensionality. To solve this problem of dimensionality, the feature selection and feature
extraction methods are used. Hereby in this paper, we are going to discuss, the different techniques for feature
extraction and feature selection and how these techniques are used to reduce the dimensionality of feature space
to improve the performance of text categorization.
Text-Image Separation in Document Images Using Boundary/Perimeter DetectionIDES Editor
Document analysis plays an important role in office
automation, especially in intelligent signal processing. The
proposed system consists of two modules: block segmentation
and block identification. In this approach, first a document is
segmented into several non-overlapping blocks by utilizing a
novel recursive segmentation technique, and then extracts
the features embedded in each segmented block are extracted.
Two kinds of features, connected components and image
boundary/perimeter features are extracted. The features are
verified to be effective in characterizing document blocks.
Last, the identification module to determine the identity of
the considered block is developed. Wide verities of documents
are used to verify the feasibility of this approach.
Content-based Image Retrieval Using The knowledge of Color, Texture in Binary...Zahra Mansoori
This document presents a new approach for content-based image retrieval that combines color, texture, and a binary tree structure to describe images and their features. Color histograms in HSV color space and wavelet texture features are extracted as low-level features. A binary tree partitions each image into regions based on color and represents higher-level spatial relationships. The performance of the proposed system is evaluated on a subset of the COREL image database and compared to the SIMPLIcity image retrieval system. Experimental results show the proposed system has better retrieval performance than SIMPLIcity in some categories and comparable performance in others.
This document summarizes an evaluation of texture feature extraction methods for content-based image retrieval, including co-occurrence matrices, Tamura features, and Gabor filters. The evaluation tested these methods on a Corel image collection using Manhattan distance as the similarity measure. Co-occurrence matrices performed best with homogeneity as the feature, while Gabor wavelets showed better performance for homogeneous textures of fixed sizes. Tamura features performed poorly with directionality. Overall, co-occurrence matrices provided the best results for general texture retrieval.
EFFECTIVE SEARCH OF COLOR-SPATIAL IMAGE USING SEMANTIC INDEXINGIJCSEA Journal
Most of the data stored in libraries are in digital form will contain either pictures or video, which is tough to search or browse. Methods which are automatic for searching picture collections made large use of color histograms, because they are very strong to wide changes in viewpoint, and can be calculated trivially. However, color histograms unable to present spatial data, and therefore tend to give lesser results. By using combination of color information with spatial layout we have developed several methods, while retrieving the advantages of histograms. A method computes a given color as a function of the distance between two pixels, which we call a color correlogram. We propose a color-based image descriptor that can be used for image indexing based on high-level semantic concepts. The descriptor is
based on Kobayashi’s Color Image Scale, which is a system that includes 130 basic colors combined in 1180 three-color combinations. The words are represented in a two dimensional semantic space into groups based on perceived similarity. The modified approach for statistical analysis of pictures involves transformations of ordinary RGB histograms. Then a semantic image descriptor is derived, containing semantic data about both color combinations and single colors in the image.
Text-Image Separation in Document Images using Boundary/Perimeter DetectionIDES Editor
Document analysis plays an important role in office
automation, especially in intelligent signal processing. The
proposed system consists of two modules: block segmentation
and block identification. In this approach, first a document is
segmented into several non-overlapping blocks by utilizing a
novel recursive segmentation technique, and then extracts
the features embedded in each segmented block are extracted.
Two kinds of features, connected components and image
boundary/perimeter features are extracted. Document with
text inside image pose limitations in earlier reported literature.
This is taken care of by applying additional pass of the Run
Length Smearing on the extracted image that contains text.
Proposed scheme is independent of type and language of the
document.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
A Novel Document Image Binarization For Optical Character RecognitionEditor IJCATR
This paper presents a technique for document image binarization that segments the foreground text accurately from poorly
degraded document images. The proposed technique is based on the
Segmentation of text from poorly degraded document images and
it is a very demanding job due to the high variation between the background and the foreground of the document. This paper pr
oposes
a novel document image binarization technique that segments t
he texts by using adaptive image contrast. It is a combination of the
local image contrast and the local image gradient that is efficient to overcome variations in text and background caused by d
ifferent
types degradation effects. In the proposed technique
, first an adaptive contrast map is constructed for a degraded input document
image. The contrast map is then binarized by global thresholding and pooled with Canny’s edge map detection to identify the t
ext
stroke edge pixels. By applying Segmentation the
text is further segmented by a local thresholding method that. The proposed method
is simple, strong, and requires minimum parameter tuning
User Interactive Color Transformation between ImagesIJMER
Abstract: In this paper we present a process called color
transfer which can borrow one image’s color
characteristics from another. Most current colorization
algorithms either require a significant user effort or have
large computational time. Here focus on orthogonal color
space i.e. lαβ color space without correlation between the
axes is given. Here we have implemented two global color
transfer algorithms in lαβ color space using simple color
statistical information such as mean, standard deviation
and covariance between the pixels of image. Our approach
is the extension of Reinhard's. Our local color transfer
algorithm uses simple color statistical analysis to recolor
the target image according to selected color range in
source image. Target image’s color influence mask is
prepared. It is a mask that specifies what parts of target
image will be affected according to selected color range.
After that target image is recolored in lαβ color space
according to prepared color influence map. In the lαβ
color space luminance and chrominance information is
separate so it allows making image recoloring optional.
The basic color transformation uses stored color statistics
of source and target image. All the algorithms are
implemented in JAVA object oriented language. The main
advantage of proposed method over the existing one is it
allows the user to recolor a part of the image in a simple &
intuitive way, preserving other color intact & achieving
natural look.
Index Terms: color transfer, local color statistics, color
characteristics, orthogonal color space, color influence
map.
CONTRAST ENHANCEMENT TECHNIQUES USING HISTOGRAM EQUALIZATION METHODS ON COLOR...IJCSEA Journal
Histogram equalization (HE) is a simple and widely used image contrast enhancement technique. The basic disadvantage of HE is it changes the brightness of the image. In order to overcome this drawback, various HE methods have been proposed. These methods preserves the brightness on the output image but, does not have a natural look. In order to overcome this problem the, present paper uses Multi-HE methods, which decompose the image into several sub images, and classical HE method is applied to each sub image. The algorithm is applied on various images and has been analysed using both objective and subjective assessment.
The document presents a novel method for character segmentation of vehicle license plates written in two rows. It discusses how the license plate image is first preprocessed through steps like grayscale conversion, binarization, and noise removal. Horizontal and vertical projections are then used to segment the image into lines and words. Character segmentation is done by analyzing the spacing between characters. Finally, zone segmentation is performed to divide each character into four zones to extract features for recognition. The method aims to simplify license plate image analysis for applications like traffic monitoring and surveillance.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
This document presents a hybrid approach for color image segmentation that integrates color edge information and seeded region growing. It uses color edge detection in CIE L*a*b color space to select initial seed regions and guide region growth. Seeded region growing is performed based on color similarity between pixels. The edge map and region map are fused to produce homogeneous regions with closed boundaries. Small regions are then merged. The approach is tested on images from the Berkeley segmentation dataset and produces reasonably good segmentation results by combining color and edge information.
This document summarizes a research paper that proposes using fuzzy logic and an artificial bee colony (ABC) optimization technique to enhance degraded color images. The approach first converts image pixels from RGB to HSV color space. It then fuzzifies the value and saturation components using Gaussian membership functions while preserving hue. The ABC optimization technique is used to optimize the membership function parameters to improve convergence time and image quality. Experimental results showed this approach yielded better results than one using ant colony optimization, converging faster while enhancing image clarity.
This document proposes a new method for segmenting outdoor images called Color Cluster Elimination (CCE) which utilizes color clustering and texture analysis. CCE performs color clustering in a multi-resolution pyramid to gradually eliminate larger color clusters, preventing them from dominating segmentation and allowing smaller clusters to emerge more clearly. It then examines regions for adjacent homochromatic objects with different textures, introducing Texture Sewn Response (TSR) to indicate texture strength across resolutions/directions. The method is evaluated on the BSDS500 dataset against other metrics, demonstrating satisfactory performance for outdoor scene segmentation.
Evaluation of Euclidean and Manhanttan Metrics In Content Based Image Retriev...IJERA Editor
This document evaluates the performance of the Euclidean and Manhattan distance metrics in a content-based image retrieval system. It finds that the Manhattan distance metric showed better precision than the Euclidean distance metric. The system uses color histograms and Gabor texture features to represent images. Color is represented in HSV color space and histograms of hue, saturation and value are used. Gabor filters are applied to capture texture at different scales and orientations. Distance between feature vectors is calculated using Euclidean and Manhattan distance formulas to find similar images from the database. The system was tested on a dataset of 1000 Corel images and Manhattan distance produced more relevant search results.
This document discusses the application of morphological image processing in forensics for fingerprint enhancement. It provides background on morphological operations like dilation, erosion, opening and closing. It explains how these operations can be used to enhance degraded fingerprints by thickening ridges, joining broken ridges, and separating overlapped ridges. The morphological image processing concepts are implemented in Java to experimentally enhance fingerprint images and reduce noise.
The document describes an algorithm for detecting text in camera-captured images. It begins with preprocessing steps like converting the color image to grayscale, applying edge detection and morphological operations like dilation and erosion. This gives initial bounding boxes containing candidate text regions. Further processing includes applying geometrical constraints to filter boxes, performing multiresolution analysis, connected component analysis and filtering by area to get the final text regions. Inversion and addition steps are used to handle text against different backgrounds.
A comparative study on content based image retrieval methodsIJLT EMAS
Content-based image retrieval (CBIR) is a method of
finding images from a huge image database according to persons’
interests. Content-based here means that the search involves
analysis the actual content present in the image. As database of
images is growing daybyday, researchers/scholars are searching
for better techniques for retrieval of images maintaining good
efficiency. This paper presents the visual features and various
ways for image retrieval from the huge image database.
Nowadays character recognition has gained lot of attention in the field of pattern recognition due to its application in various fields. It is one of the most successful applications of automatic pattern recognition. Research in OCR is popular for its application potential in banks, post offices, office automation etc. HCR is useful in cheque processing in banks; almost all kind of form processing systems, handwritten postal address resolution and many more. This paper presents a simple and efficient approach for the implementation of OCR and translation of scanned images of printed text into machine-encoded text. It makes use of different image analysis phases followed by image detection via pre-processing and post-processing. This paper also describes scanning the entire document (same as the segmentation in our case) and recognizing individual characters from image irrespective of their position, size and various font styles and it deals with recognition of the symbols from English language, which is internationally accepted.
SEGMENTATION OF CHARACTERS WITHOUT MODIFIERS FROM A PRINTED BANGLA TEXTcscpconf
This document discusses the segmentation of printed Bangla characters without modifiers for optical character recognition systems. It begins with an introduction to OCR systems and Bangla script. The main steps of an OCR system are then outlined, with a focus on the segmentation step. Line, word and character segmentation algorithms are described in detail along with figures to illustrate the steps. The goal is to properly segment individual characters for recognition.
This document describes an extended fuzzy c-means (EFCM) clustering algorithm for noisy image segmentation. The algorithm first preprocesses noisy pixels in an image by regenerating their values based on neighboring pixel intensities. It then applies the conventional fuzzy c-means clustering algorithm to segment the image. The EFCM approach is presented as being less sensitive to noise than other clustering algorithms and able to efficiently segment noisy images. The document provides background on image segmentation, fuzzy c-means clustering, types of image noise, and density-based clustering challenges. It also outlines the EFCM methodology and its computational advantages over other robust clustering methods for noisy image segmentation.
Feature Extraction and Feature Selection using Textual Analysisvivatechijri
After pre-processing the images in character recognition systems, the images are segmented based on
certain characteristics known as “features”. The feature space identified for character recognition is however
ranging across a huge dimensionality. To solve this problem of dimensionality, the feature selection and feature
extraction methods are used. Hereby in this paper, we are going to discuss, the different techniques for feature
extraction and feature selection and how these techniques are used to reduce the dimensionality of feature space
to improve the performance of text categorization.
A Review of Optical Character Recognition System for Recognition of Printed Textiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
The document summarizes the key steps in an optical character recognition (OCR) system for recognizing printed text:
1. Image acquisition involves obtaining the image, which can be done using scanners or digital cameras.
2. Pre-processing prepares the image for recognition through techniques like converting to grayscale, skew correction, binarization, noise reduction, and thinning.
3. Segmentation separates the image into lines and individual characters.
4. Recognition identifies the characters by comparing features or templates to stored models.
The paper then discusses specific algorithms that could implement grayscale conversion, skew correction, and other steps in the OCR system.
The document discusses color-based video retrieval using block and global methods. It summarizes that color features are widely used in video retrieval and content analysis. It describes extracting color histograms globally and from divided blocks. Two methods are discussed: global color extracts overall color frequency while block color quantizes each divided region, maintaining some spatial data. Videos are retrieved by comparing feature vectors of queries to those in a database using distance metrics like Euclidean distance. MATLAB is used to implement the color feature extraction and retrieval methods.
Color is a widely used visual feature for content-based video retrieval. There are two main methods discussed in the document: block-based and global color feature extraction. The block-based method extracts color histograms from divided blocks of each video frame, while the global method extracts a single color histogram from the entire frame. These color features are then used to measure similarity between videos for retrieval. The document also discusses challenges with high-dimensional color histograms and methods to reduce dimensions like transforms and selecting significant colors. Overall the paper presents color-based video retrieval techniques and evaluates performance of the block-based and global methods.
The document presents a hybrid approach for detecting and recognizing text in images. It consists of three main steps:
1) Image partition using k-means clustering to segment text regions based on color information.
2) Character grouping to detect text characters within each text string based on character size differences and distance between characters.
3) Text recognition of detected characters using a neural network.
The proposed method was evaluated on a street view text dataset, achieving a precision of 0.83, recall of 0.93, and f-measure of 0.25 for text recognition. The approach efficiently and accurately detects and recognizes text with low false positives.
International Journal of Computational Engineering Research(IJCER) ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
The technical study had been performed on
many foreign languages like Japanese; Chinese etc. but the
efforts on Indian ancient script is still immature. As the Modi
script language is ancient and cursive type, the OCR of it is still
not widely available. As per our knowledge, Prof. D.N.Besekar,
Dept. of Computer Science, Shri. Shivaji College of Science,
Akola had proposed a system for recognition of offline
handwritten MODI script Vowels. The challenges of
recognition of handwritten Modi characters are very high due
to the varying writing style of each individual. Many vital
documents with precious information have been written in
Modi and currently, these documents have been stored and
preserved in temples and museums. Over a period of time these
documents will wither away if not given due attention. In this
paper we propose a system for recognition of handwritten
Modi script characters; the proposed method uses Image
processing techniques and algorithms which are described
below.
General Terms
Preprocessing techniques: Gray scaling, Thresholding,
Boundary detection, Thinning, cropping, scaling, Template
generation. Other algorithms used- Average method, otsu
method, Stentiford method, Template-based matching method
OCR for Gujarati Numeral using Neural Networkijsrd.com
This papers functions within to reduce individuality popularity (OCR) program for hand-written Gujarati research. One can find so much of work for Indian own native different languages like Hindi, Gujarati, Tamil, Bengali, Malayalam, Gurumukhi etc., but Gujarati is a vocabulary for which hardly any work is traceable especially for hand-written individuals. Here in this work a nerve program is provided for Gujarati hand-written research popularity. This paper deals with an optical character recognition (OCR) system for handwritten Gujarati numbers. A several break up food ahead nerve program is suggested for variation of research. The functions of Gujarati research are abstracted by four different details of research. Reduction and skew- changes are also done for preprocessing of hand-written research before their variation. This work has purchased approximately 81% of performance for Gujarati handwritten numerals.
The document summarizes an automatic text extraction system for complex images. The system uses discrete wavelet transform for text localization. Morphological operations like erosion and dilation are used to enhance text identification and segmentation. Text regions are segmented using connected component analysis and properties like area and bounding box shape. The extracted text is recognized and shown in a text file. The system allows modifying the recognized text and shows better performance than existing techniques.
Multimedia Big Data Management Processing And AnalysisLindsey Campbell
The document discusses various techniques used for extracting useful information from images, including image classification, feature extraction, face detection and recognition, and image retrieval. Image classification involves applying machine learning algorithms to assign images to predefined categories or classes. Feature extraction identifies distinguishing aspects of images like color, shape and texture. Face detection locates human faces within images while face recognition identifies specific faces by comparing features to face databases. Image retrieval finds similar images in databases based on visual features. These techniques extract meaningful information from images to enable enhanced image searching capabilities.
IRJET - An Enhanced Approach for Extraction of Text from an Image using Fuzzy...IRJET Journal
This document presents an approach for extracting text from images using fuzzy logic. It involves preprocessing the image to remove noise, segmenting the image to extract individual characters, and then using fuzzy logic to identify the characters by comparing segmented characters to trained data and determining the degree of matching. The key steps are pre-processing, segmentation, feature extraction using techniques like statistical and geometrical features, classification using a convolutional neural network, and then using fuzzy logic to accurately identify characters by finding the highest matching value between segmented and trained characters. The goal is to recognize and extract text from the image in an editable format.
A novel embedded hybrid thinning algorithm forprjpublications
The document proposes a hybrid thinning algorithm that combines the Stentiford and Zhang-Suen thinning algorithms. It compares the hybrid algorithm to the original Stentiford and Zhang-Suen algorithms on an input image. The hybrid algorithm more accurately thins the image to a single pixel width but does not improve time complexity compared to the original algorithms. The hybrid approach uses four templates across two sub-iterations to identify and remove pixels based on connectivity values until no more can be removed. Experimental results show the hybrid algorithm more effectively increases image contrast than the original thinning algorithms.
This document summarizes image indexing and its features. It discusses that image indexing is used to retrieve similar images from a database based on extracted features like color, shape, and texture. Color features can be represented by models like RGB, HSV, and color histograms. Shape features include global properties like roundness and local features like edge segments. Texture is described using statistical, structural, and spectral approaches. Texture feature extraction methods discussed include standard wavelets, Gabor wavelets, and extracting features like entropy and standard deviation. The paper provides an overview of the different features used for image indexing and classification.
The document discusses image processing and provides information on several key topics:
1. Image processing can be grouped into compression, preprocessing, and analysis. Preprocessing improves image quality by reducing noise and enhancing edges. Analysis extracts numeric or graphical information for tasks like classification.
2. Images are 2D matrices of intensity values represented by pixels. Common digital formats include grayscale, RGB, and RGBA. Higher bit depths allow more intensity levels to be represented.
3. Basic measurements of images include spatial resolution in pixels per unit, bit depth determining representable intensity levels, and factors like saturation and noise.
An Unsupervised Cluster-based Image Retrieval Algorithm using Relevance FeedbackIJMIT JOURNAL
Content-based image retrieval (CBIR) systems utilize low level query image feature as identifying similarity between a query image and the image database. Image contents are plays significant role for image retrieval. There are three fundamental bases for content-based image retrieval, i.e. visual feature extraction, multidimensional indexing, and retrieval system design. Each image has three contents such as: color, texture and shape features. Color and texture both plays important image visual features used in Content-Based Image Retrieval to improve results. Color histogram and texture features have potential to retrieve similar images on the basis of their properties. As the feature extracted from a query is low level, it is extremely difficult for user to provide an appropriate example in based query. To overcome these problems and reach higher accuracy in CBIR system, providing user with relevance feedback is famous for provide promising solutio
For electrically powered vehicles Battery Management Systems are responsible for protected
charging of the batteries. Different switches like MOSFETS, transistors are used to control the
charging current of the batteries. While charging the batteries with 48A/11KW or 16A/45 KW the
switches generate heat up to 150 W. The air environment in parked vehicles is not sufficient to cool
the BMS. As a result the temperature rises up to 230°C, which in turn can change the nominal
value of different components like resistors, semiconductor switches, inductances, capacitances,
and the system could behave abnormally. Therefore, a reliable system must be appointed here to
cool the BMS. The main goal of this paper is to propose and develop a system which can be a
perfect choice for BMS cooling. There are many cooling systems which can be considered for this
purpose. However, the Thermoelectric Cooler (TEC): Peltier cooler is the best option, which we
have analysed from different point of views. The solution approaches included for evaluation are a
physical model proposition, from that model designing a MATLAB simulator and after that various
perspectives are elucidated using this simulator. Another important utilisation of the MATLAB
simulator is that any TEC system independent of sizes can be simulated here and the performance
of the system can also be realised visually
Comparative study of machine learning algorithms for wind speed prediction in...Mohammad Liton Hossain
This study evaluated the performance of multiple models that used machine learning to anticipate wind speed
in the city of Dhaka. The NASA Power website provided the data set for this investigation. The models used for prediction included the decision tree regressor, support vector regressor, random forest, linear regression, neural network
and polynomial regression. A hold-out check and k-fold cross-validation were used to assess how well these models
performed. With the highest R2 scores and lowest RMSEs on both the validation and test sets, the results demonstrated that the polynomial regression model performed the best. With the lowest R2 scores and largest RMSEs
on both sets, the decision tree model scored the poorest. High R2 scores and low RMSEs were achieved by the random forest model, which had comparable performance to the polynomial regression model but required a longer
computation time. In addition, the neural network model demonstrated commendable predictive accuracy, yielding
an R2 score of 0.67 and a low RMSE of 0.57. However, its application is contingent on the availability of substantial
computational resources, given its extensive computation time of 457.93 s. The study concludes by highlighting
the efficacy of the Polynomial Regression model as the optimal choice for wind speed prediction in Dhaka, offering
a balance between superior performance and efficient computation. This insight provides valuable guidance for practitioners and researchers seeking effective models for similar applications.
Harnessing the power of wind: a comprehensive analysis of wind energy potenti...Mohammad Liton Hossain
This study gives a thorough analysis on the wind energy potential in Dhaka, Bangladesh, utilizing data from NASA
Power’s remote sensing tools and weather data from the Bangladesh Meteorological Department (BMD). The wind
speed data collected over a 22 year period at an altitude of 10 m. The results indicate that 3.07 m per second (ms−1)
is Dhaka city’s typical wind speed, while the maximum wind speeds were recorded in June and July. A Weibull
distribution is used to observe the wind data, as well as to calculate the Weibull form parameter of 2.65 and the scale
parameter of 3.43 ms−1. Based on these parameters, the most probable wind speed along with the wind speed
carrying maximum energy were calculated 2.83 ms−1 and 4.28 ms−1, respectively. The highest density of energy
has been found in the month of July with a value of 52.11 W/m2. According to the study, the south is the most
prominent wind direction for Dhaka city. Moreover, the study analyzes the relations between energy density
and other variables, like wind speed, humidity, dry bulb temperature, etc. Positive correlations between energy
density, wind speed, and dry bulb temperature imply that the higher wind speeds and dry bulb temperatures result
in greater energies. The study’s conclusions offer intuitive information about Dhaka City’s potential for wind energy
and can support direct future efforts to pursue this green resource in alignment with the Sustainable Development
Goals (SDGs) of Bangladesh
Developing a hands-free interface to operate a Computer using voice commandMohammad Liton Hossain
The main focus of this study is to help a handicap person to operate a computer by voice command. It can be used to operate the entire computer functions on the user’s voice commands. It makes use of the Speech Recognition technology that allows the computer system to identify and recognize words spoken by a human using a microphone. This Software will be able to recognize spoken words and enable user to interact with the computer. This interaction includes user giving commands to his computer which will then respond by performing several tasks, actions or operations depending on the commands they gave. For Example: Opening /closing a file in computer, YouTube automation using voice command, Google search using voice command, make a note using voice command, calculation by calculator using voice command etc.
The main focus of this study is to find appropriate and stable solutions for representing the statistical data into map with some special features. This research also includes the comparison between different solutions for specific features. In this research I have found three solutions using three different technologies namely Oracle MapViewer, QGIS and AnyMap which are different solutions with different specialties. Each solution has its own specialty so we can choose any solution for representing the statistical data into maps depending on our criteria’s.
Development of an Audio Transmission System Through an Indoor Visible Light ...Mohammad Liton Hossain
This study presents an approach to develop an indoor visible light communication system capable of transmitting audio signal over light beam within a short distance. Visible Light Communication (VLC) is a pretty new technology which used light sources to transmit data for communication. In any communication system, both analog and digital signal transmission are possible, though, due to having the capability of providing a faithful quality of signal regeneration after the transmission process, digital communication system is much more popular than the analog one. In the current project, digital communication process was adopted also. To convert the analog audio signal into the digital transmission signal and vice versa, Pulse Width Modulation (PWM) was used as the signal encoding strategy. As the light emitter, white Light Emitting Diodes (LEDs) were used and as photo sensor, a solar cell was used instead of a photodiode to obtain greater signal power and sensitivity. In the system, the carrier signal for transmission was chosen to have a frequency of 50 KHz. At the receiving end, a 4th order Butterworth lowpass filter having a cutoff frequency of 8 KHz was used to demodulate the audio signal. Using only 2 white LEDs, the indoor transmission range of this visible light communication system was found to be 5 meters while reproducing a satisfactory quality audio.
Development of a Low Power Indoor Transmission System with a Dedicated Androi...Mohammad Liton Hossain
This document describes the development of a low power indoor transmission system using a Raspberry Pi with an attached antenna. The system is designed to function as a campus radio that can broadcast educational programs and updates to students. Key aspects include using the Raspberry Pi's built-in FM transmitter capabilities along with a custom antenna. In addition, an Android app was developed to allow remote management of broadcast scheduling and content uploading. The system was tested successfully and could provide an alternative information source for students if implemented on a larger scale at universities.
DESIGN AND IMPLEMENTATION A BPSK MODEM AND BER MEASUREMENT IN AWGN CHANNELMohammad Liton Hossain
This document describes the design and implementation of a BPSK modem and the measurement of bit error rate (BER) in an additive white Gaussian noise (AWGN) channel. It includes:
1) The design of a transmitter that generates a binary data sequence, modulates it using BPSK, and transmits it through an AWGN channel.
2) The design of a receiver that recovers the transmitted data by demodulating the noisy received signal using correlation and threshold detection.
3) The use of MATLAB to simulate the designed BPSK modem and measure BER at different signal-to-noise ratios (SNR) to evaluate the system performance. The measured BER is also compared to theoretical BER estimates.
The main focus of this study is to find appropriate and stable solutions for representing the statistical data into map with some special features. This research also includes the comparison between different solutions for specific features. In this research I have found three solutions using three different technologies namely Oracle MapViewer, QGIS and AnyMap which are different solutions with different specialties. Each solution has its own specialty so we can choose any solution for representing the statistical data into maps depending on our criteria’s.
MODULE 5 BUILDING PLANNING AND DESIGN SY BTECH ACOUSTICS SYSTEM IN BUILDINGDr. BASWESHWAR JIRWANKAR
: Introduction to Acoustics & Green Building -
Absorption of sound, various materials, Sabine’s formula, optimum reverberation time, conditions for good acoustics Sound insulation:
Acceptable noise levels, noise prevention at its source, transmission of noise, Noise control-general considerations
Green Building: Concept, Principles, Materials, Characteristics, Applications
This presentation provides a comprehensive overview of a specialized test rig designed in accordance with ISO 4548-7, the international standard for evaluating the vibration fatigue resistance of full-flow lubricating oil filters used in internal combustion engines.
Key features include:
Filters for Electromagnetic Compatibility ApplicationsMathias Magdowski
In this lecture, I explain the fundamentals of electromagnetic compatibility (EMC), the basic coupling model and coupling paths via cables, electric fields, magnetic fields and wave fields. We also look at electric vehicles as an example of systems with many conducted EMC problems due to power electronic devices such as rectifiers and inverters with non-linear components such as diodes and fast switching components such as MOSFETs or IGBTs. After a brief review of circuit analysis fundamentals and an experimental investigation of the frequency-dependent impedance of resistors, capacitors and inductors, we look at a simple low-pass filter. The input impedance from both sides as well as the transfer function are measured.
ISO 4020-6.1 – Filter Cleanliness Test Rig: Precision Testing for Fuel Filter Integrity
Explore the design, functionality, and standards compliance of our advanced Filter Cleanliness Test Rig developed according to ISO 4020-6.1. This rig is engineered to evaluate fuel filter cleanliness levels with high accuracy and repeatability—critical for ensuring the performance and durability of fuel systems.
🔬 Inside This Presentation:
Overview of ISO 4020-6.1 testing protocols
Rig components and schematic layout
Test methodology and data acquisition
Applications in automotive and industrial filtration
Key benefits: accuracy, reliability, compliance
Perfect for R&D engineers, quality assurance teams, and lab technicians focused on filtration performance and standard compliance.
🛠️ Ensure Filter Cleanliness — Validate with Confidence.
Module4: Ventilation
Definition, necessity of ventilation, functional requirements, various system & selection criteria.
Air conditioning: Purpose, classification, principles, various systems
Thermal Insulation: General concept, Principles, Materials, Methods, Computation of Heat loss & heat gain in Buildings
Video Games and Artificial-Realities.pptxHadiBadri1
🕹️ #GameDevs, #AIteams, #DesignStudios — I’d love for you to check it out.
This is where play meets precision. Let’s break the fourth wall of slides, together.
This presentation provides a detailed overview of air filter testing equipment, including its types, working principles, and industrial applications. Learn about key performance indicators such as filtration efficiency, pressure drop, and particulate holding capacity. The slides highlight standard testing methods (e.g., ISO 16890, EN 1822, ASHRAE 52.2), equipment configurations (such as aerosol generators, particle counters, and test ducts), and the role of automation and data logging in modern systems. Ideal for engineers, quality assurance professionals, and researchers involved in HVAC, automotive, cleanroom, or industrial filtration systems.
All about the Snail Power Catalog Product 2025kstgroupvn
DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BANGLADESH
1. International Journal of Scientific and Research Publications, Volume 9, Issue 1, January 2019 228
ISSN 2250-3153
http://dx.doi.org/10.29322/IJSRP.9.01.2019.p8530 www.ijsrp.or
Development of an Alphabetic Character Recognition
System Using Matlab for Bangladesh
Mohammad Liton Hossain*
, Tafiq Ahmed**, S.Sarkar***
, Md. Al-Amin****
*
Department of ECE, Institute of Science and Technology, National University
**
Electrical Engineering, University of Rostock, Germany
DOI: 10.29322/IJSRP.9.01.2019.p8530
http://dx.doi.org/10.29322/IJSRP.9.01.2019.p8530
Abstract- Character recognition technique, associates a symbolic
identity with the image of the character, is an important area in
pattern recognition and image processing. The principal idea here
is to convert raw images (scanned from document, typed,
pictured etcetera) into editable text like html, doc, txt or other
formats. There is a very limited number of Bangla Character
recognition system, if available they can’t recognize the whole
alphabet set. Motivated by this, this paper demonstrates a
MATLAB based Character Recognition system from printed
Bangla writings. It can also compare the characters of one image
file to another one. Processing steps here involved binarization,
noise removal and segmentation in various levels, features
extraction and recognition.
Index Terms- OCR, Character Recognition, MATLAB, Cross-
Correlation, ImageProcessing.
I. INTRODUCTION
Character Recognition began as a field of research in pattern
identification, artificial neural networks and machine learning.
The different areas covered under this general term are either on-
line or off-line CR, each having its dedicated hardware and
recognition methods. In on-line character recognition
applications, the computer recognizes the symbols as they are
drawn. The typical hardware for data acquisition is the digitizing
tablet, which can be electromagnetic, electrostatic, pressure
sensitive etcetera; a light pen can also be used. As the character
is drawn, the successive positions of the pen are memorized (the
usual sampling frequencies lie between 100Hz and 200Hz) and
are used by the recognition algorithm [1]. Off-line character
recognition is completed afterward the inscription or printing is
performed. Examples of it are usually distinguished: magnetic
and optical character recognition. In magnetic character
recognition (MCR), the fonts are written with magnetic disk and
are designed to modify in a unique way a magnetic field created
by the acquisition device. MCR is mostly used in banking
applications, for instance reading bank checks, because
overwriting or overprinting these characters does not affect the
accuracy of the recognition. While in optical character
recognition, which is the field investigated in this paper, deals
with recognition of characters acquired by optical means,
typically by a scanner or a camera. The symbols can be separated
from each other or belong to structures like words, paragraphs,
figures, etc. They can be printed or handwritten, of any size,
shape, or orientation [2, 3]. Bangla, the mother tongue of
Bangladeshis, is one of the most popular languages in the world,
For the Indian subcontinent Bangla is the Second most popular
language. Moreover 200 million people of eastern India and
Bangladesh use this language and also the institutional language
in our country [1]. So recognition of Bangla alphabet is always a
special interest to us considering the massive number of official
papers are being scanned and processed every day. This work
aims to develop a system that categorizes a given input (Bangla
Characters) as belonging to a certain class rather than to identify
them uniquely, as every input pattern. It also performs character
recognition by quantification of the character into a mathematical
vector entity using the geometrical properties of the character
image. Global usage of this system may reduce the hard labor of
the concerning government employees every day.
II. BACKGROUND
Basic Bangla character set comprises 11 vowels, 39 consonant
and 10 numerals. There are also compound characters being
combination of consonant with consonant as well as consonant
with vowel [1, 4]. The complete set of characters, need to
recognize by the proposed recognition system, is provided in
Figure1.
Figure1: The whole character set of Bangla Alphabet
A. OVERVIEW OF THE IMAGE TYPES AND
PROPERTIES
The RGB (true-color) images are put in storage as 3 distinct
image matrices; one of red (R) in each pixel, one of green (G)
and one of blue (B); similar to the wavelength-sensitive cone
cells of human eye. While in a grayscale or graylevel digital
image, the value of each pixel is a single value, carries only
intensity information. A grayscale image is not the black and
white Binary (2-leveled) image. Gray image is represented by
black and white shades or combination of levels. 8 bit gray image
2. International Journal of Scientific and Research Publications, Volume 9, Issue 1, January 2019 229
ISSN 2250-3153
http://dx.doi.org/10.29322/IJSRP.9.01.2019.p8530 www.ijsrp.or
represents total 2^8 levels, from black to white, 0 equals black
and 255 is White. While Binary images, have only two probable
values for individual pixel, are produced by thresholding a
grayscale or color image to separate an object in the image from
its background [5].
Also, color images are generally described by using the
following three terms: Hue, Saturation and Lightness. Hue is
another word for color and dependent on the wavelength of light
being reflected or produced. Saturation refers to that how pure
or intense a given hue is. 100% saturation means there’s no
addition of gray to the hue - the color is completely pure [15].
Lightness measures the relative degree of black or white that’s
been mixed with a given hue. Adding white makes the color
lighter (creates tints) and adding black makes it darker (creates
shades). The effect of lightness is relative to other values in the
composition [6, 15].
III. METHODOLOGY
The proposed Bangla Chracter Recognition system is
developed here by the following steps:
Step1: Printed Bangla Script input to Scanner
Step2: Raw scanned Document from Scanner
Step3: Convert the scanned RGB image into Grayscale
image
Step3: Convert Grayscale image to Binary image.
Step4: Segmentation and Feature Extraction.
Step5: Classification and Post-processing.
Step6: Output editable text.
A scanned RGB color image here is input to the system. The
proposed text detection, recognition and translation algorithm
(shown in Figure2) consists of following steps: Binarization,
Noise removal, Segmentation, Features extraction, Correlation
and Cross-Correlation [1, 7].
A. Pre-Processing
Binarization
Binarization is the technique here by which the gray scale
pictures are transformed to binary images to facilitate noise
removal. Typically the two colors used for a binary image are
black and white though any two colors can be used. The color
used for the object in the image is the foreground color while the
rest of the image is the background color . Binarization separates
the foreground (text) from the background information [1-2].
Figure 2: An overview of the proposed Bangla character recognition
system.
Noise Removal
Scanned documents often contain noise that arises due to the
accessories of printer or scanner, print quality, age of the
document, etc. Therefore, it is necessary to filter this noise before
processing the image. This low-pass filter should avoid as much
of the distortion as possible while holding the entire signal [1, 8].
B. Segmentation
Image segmentation is the process of partitioning a digital image
into multiple segments (sets of pixels, also known as super
pixels). The goal of segmentation is to simplify and/or change
the representation of an image into something that is more
meaningful and easier to analyze. Image segmentation is the
process of assigning a label to every pixel in an image such that
pixels with the same label share certain characteristics or
property such as color, concentration, or quality. The result of
image segmentation is a set of sections that collectively cover the
whole image, or a set of lines take out from the image (or edge
detection). Segmentation of binary image is achieved in altered
levels contains line segmentation, word segmentation, character
segmentation. Thresholding often delivers an easy and suitable
method to accomplish this segmentation on the base of the
different concentrations or colors in the foreground and
background areas. In a single pass, each pixel in the image is
related with this threshold value. If the pixel's intensity is higher
than the threshold, the pixel is set to white in the output,
corresponding foreground. If it is less than the threshold, it is set
to black, indicating background [1, 2, 8].
Features Extraction
Feature taking out is the special form of dimensional decrease
that proficiently characterizes exciting parts of an image as a
solid feature route. This method is suitable when image sizes are
huge and a condensed feature illustration is essential to swiftly
complete tasks, such as image matching and recovery. Feature
detection, feature mining, and matching are often combined to
solve common computer vision problems such as object
detection and recognition, content-based image retrieval, face
detection and recognition, and texture classification [1, 9].
3. International Journal of Scientific and Research Publications, Volume 9, Issue 1, January 2019 230
ISSN 2250-3153
http://dx.doi.org/10.29322/IJSRP.9.01.2019.p8530 www.ijsrp.or
C. Correlation and Cross-Corellation
Image correlation and tracking is an optical technique that works
tracking and image registration method for accurate 2D and 3D
dimension of changes in images. Cross-correlation is the
measure of similarity of two images. Each of the images is
divided into rectangular blocks. Each block in the first image is
correlated with its corresponding block in the second image to
produce the cross-correlation value as a function of position
[10,11].
IV. MATLAB IMPLEMENTATION
An implementation is the realization of a technical specification
or algorithm as a program, software component, or other
computer system through computer programming and
deployment. The proposed system is implemented in MATLAB
and the accuracy is also tested. Using MATLAB ‘imread’
function of, a noisy scanned RGB image is loaded to the input
system (Figure.3). After performing Grayscale conversion by
‘rgb2gray’, Figure.4 is thus obtained; ‘rgb2gray’ eliminates the
hue and saturation information of the RGB image while retaining
the luminance. Figure.5-6 shows outcome after binarization
(using ‘im2bw’) and noise-removal from this black and white
image. Figure.7-8 shows the process of character segmentation
[12-13].
Performing cross-correlation between input image and template
image, the peak point is found (Figure.9). Peak point is the point
that indicates the highest matching area based on cross-
correlation. Thus, the appropriate character is identified
(Figure.10). Here Bangla Unicode Characters’ hex value is used
to make the recognized character machine-readable. Recognized
character is here displayed in MATLAB built in browser [14]. Figure 9: Marked Peak in cross correlation refer Character Detection
Figure 10: Recognized character from Template.
Figure 7: Image information for segmentation Figure 8: Character after Segmentation.
0 represented by Black & 1 represents White.
4. International Journal of Scientific and Research Publications, Volume 9, Issue 1, January 2019 231
ISSN 2250-3153
http://dx.doi.org/10.29322/IJSRP.9.01.2019.p8530 www.ijsrp.or
V. GUI (GRAPHICAL USER INTERFACE)
DEVELOPMENT
A user-friendly Graphical User Interface is also designed for this
system, where user can insert an image to recognize the desired
character. User can input any image using “Load Image” option.
After loading an image, user can select a character using mouse
pointer, crop image and continue processing using “Image
Processing” option. Following the algorithms, the system would
recognize the selected character and provide it in editable format
if “Recognize Character” command is executed from the options.
VI. ANALYSIS
The recognition rate of the proposed recognition method is
remarkable with the accuracy rate of almost 98-100%. Image
processing toolbox and built in MATLAB functions have been
quite helpful here for detection of single Bangla character and
digit. The input images can be taken from any optical scanner or
camera. The output allows saving recognized character as html,
txt etcetera format for further processing. The only limitation of
the system is the fixed font size of the character. Further
improvement for all possible font type and size is reserved for
the future. Also it is developed only for single Bangla character
recognition at a time. Multiple characters recognition along with
connected letters is also left open for the future. Connecting the
program algorithm with the neural network may also be helpful.
After the successful review and payment, IJSRP will publish
your paper for the current edition. You can find the payment
details at: http://ijsrp.org/online-publication-charge.html.
VII. CONCLUSION
Complex character based language like Bangla clearly demands
deep research to meet its goals. Recognition of Bangla characters
is still in the intermediate stage. The proposed system is
developed here using template matching approach to recognize
individual character image. Even though the user-friendly GUI
gives several advantages to individual users, this system is still
facing a number of limitations which is quite considerable.
Improving the system to suit handwritten characters, all printed
fonts in considerable amount of time with high accuracy is the
recommended future work.
ACKNOWLEDGMENT
At first all thanks goes to almighty creator who gives me the
opportunity, patients and energy to complete this study. I would
like to give thanks to Mr. Masudul Haider Imtiaz; Assistant
Professor at Institute of Science and Technology. I have found
always immense support from him to keep my work on the right
way. His door was always open for me, when I have got myself
in trouble.
Finally, I must express my very profound gratefulness to my
Figure 11: Computer editable Character
Figure 12: A specific instance of character recognition using
developed MATLAB GUI
5. International Journal of Scientific and Research Publications, Volume 9, Issue 1, January 2019 232
ISSN 2250-3153
http://dx.doi.org/10.29322/IJSRP.9.01.2019.p8530 www.ijsrp.or
parents and to my wife for providing me with constant support
and encouragement during my years of study and through the
process of researching and writing this paper. This
accomplishment would not have been possible without them.
Thank you.
REFERENCES
REFERENCES
[1] Md. Mahbub Alam and Dr. M. Abul Kashem, A Complete Bangla
OCR System for Printed Characters, JCIT, 01(01).
[2] MCR, OCR, Binary Image, Image Segmentation, Feature-extraction,
Wikipedia, the free encyclopedia.
[3] Sandeep Tiwari, Shivangi Mishra, Priyank Bhatia, Praveen Km.
Yadav, Optical Character Recognition using MATLAB, International
Journal of Advanced Research in Electronics and Communication
Engineering (IJARECE) ,5(2), (2013).
[4] Bengali Alphabet, Wikipedia, the free encyclopedia.
[5] RGB Color Model, Grayscale, Wikipedia, the free encyclopedia.
[6] Hue, saturation and brightness, Wikipedia, the free encyclopedia.
[7] Md. Alamgir Badsha, Md. Akkas Ali, Dr. Kaushik Deb, Md.
Nuruzzaman Bhuiyan, Handwritten Bangla Character Recognition
Using Neural Network, International Journal of Advanced Research in
Computer Science and Software Engineering, 11(2), (2012).
[8] Samiur Rahman Arif, Bangla Character Recognition using Feature
extraction, Project work, Department of Computer Science, Brac
University, (2007).
[9] Ahmed Asif Chowdhury et.al Optical Character Recognition of
Bangla Characters using neural network: A better approach,
Department of Computer Science, BUET, Bangladesh.
[10] Amin Ahsan Ali & Syed Monowar Hossain, Optical Bangla Digit
Recognition using Back propagation Neural Networks, Project work,
Department of computer science, University of Dhaka, (2003).
[11] Definition, Cross-correlation, Online.
[12] MATLAB, Wikipedia, the free encyclopedia.
[13] MATLAB, Image processing toolbox, Mathworks, Online.
[14] Rafael C. Gonzalez,Richard E. Woods, Setven L. Eddins, Digital
Image Processing using MATLAB, 2nd Edition, (2009).
[15] The Fundamentals of Color: Hue, Saturation, And Lightness,
available at online: https://vanseodesign.com/web-design/hue-
saturation-and-lightness/
AUTHORS
First Author –– Mohammad Liton Hossain, Lecturer,
Department of ECE, IST
[email protected]
+8801768346307
Second Author -Tafiq Ahmed, Electrical Engineering, University of
Rostock, Germany
Third Author –S.Sarkar, Department of ECE, IST