The Optical Character Recognition (OCR) is becoming
popular areas of research under pattern
recognition and smart device applications. It requi
res the intelligence like human brain to
recognize the various handwritten characters. Artif
icial Neural Network (ANN) is used to
gather the information required to recognize the ch
aracters adaptively. This paper presents a
performance analysis of character recognition by tw
o different methods (1) compressed Lower
Dimension Feature(LDF) matrix with a perceptron net
work, (2) Scale Invariant Feature (SIF)
matrix with a Back Propagation Neural network (BPN)
. A GUI based OCR system is developed
using Matlab. The results are shown for the English
alphabets and numeric. This is observed
that the perceptron network converges faster, where
as the BPN can handle the complex script
recognition when the training set is enriched.
Inpainting scheme for text in video a surveyeSAT Journals
This document summarizes text detection and removal schemes for video sequences. It discusses two main phases - text detection and video inpainting. For text detection, it describes various visual feature extraction techniques like edge detection and texture analysis. It also discusses machine learning approaches like multilayer perceptrons and support vector machines. For inpainting, it discusses using wavelet transforms to approximate boundary data and fill in missing regions after text is removed. The goal is to restore occluded parts of video frames while maintaining spatial and temporal consistency.
Segmentation and recognition of handwritten digit numeral string using a mult...ijfcstjournal
In this paper, the use of Multi-Layer Perceptron (MLP) Neural Network model is proposed for recognizing
unconstrained offline handwritten Numeral strings. The Numeral strings are segmented and isolated
numerals are obtained using a connected component labeling (CCL) algorithm approach. The structural
part of the models has been modeled using a Multilayer Perceptron Neural Network. This paper also
presents a new technique to remove slope and slant from handwritten numeral string and to normalize the
size of text images and classify with supervised learning methods. Experimental results on a database of
102 numeral string patterns written by 3 different people show that a recognition rate of 99.7% is obtained
on independent digits contained in the numeral string of digits includes both the skewed and slant data.
Character Recognition (Devanagari Script)IJERA Editor
This document summarizes research on using neural networks for optical character recognition of Devanagari script characters. It describes preprocessing scanned images, extracting features using neural networks, and post-processing to recognize characters. The system was tested on a dataset of Devanagari characters with neural networks trained over multiple epochs. Recognition accuracy increased with larger training sets as the network learned to identify characters more precisely. The system demonstrates an effective approach for digitally recognizing handwritten Devanagari characters.
This document presents a novel method for recognizing two-dimensional QR barcodes using texture feature analysis and neural networks. It first extracts texture features like mean, standard deviation, smoothness, skewness and entropy from divided blocks of barcode images. These features are then used to train a neural network to classify blocks as containing a barcode or not. The trained neural network can then be used to locate barcodes in unknown images by classifying each block. The method is implemented and evaluated using MATLAB on a database of QR code images, showing satisfactory recognition results.
HANDWRITTEN DIGIT RECOGNITION USING k-NN CLASSIFIERvineet raj
This document proposes using a k-nearest neighbor classifier to recognize handwritten digits from the MNIST database. It discusses existing methods that use star-layered histogram feature extraction and class-dependent feature selection, which achieve accuracies of around 93% and 92% respectively. However, these methods require thinning operations or have high computational costs. The document proposes using k-NN classification with pre-processing and feature extraction to achieve higher accuracy of around 96% with lower computation requirements than existing models.
Recognition Of Handwritten Meitei Mayek Script Based On Texture Featureijnlc
This document presents a method for recognizing handwritten characters of the Meitei Mayek script using texture features and a support vector machine classifier. A dataset of 3,780 handwritten characters collected from 70 people was used to develop and evaluate the recognition model. Local binary patterns were extracted as texture features from the pre-processed images. Using this approach, the highest recognition rate achieved on the dataset was 93.33%. This represents an improvement over previous work on recognizing handwritten Meitei Mayek characters. Future work could focus on developing models to recognize complete sentences instead of isolated characters.
Representation and recognition of handwirten digits using deformable templatesAhmed Abd-Elwasaa
This document presents a case study on using deformable templates for recognizing handwritten digits. Deformable templates match unknown images to known templates by deforming the contours of templates to fit the edge strengths of unknown images. The dissimilarity measure is derived from the deformation needed for the match. This technique achieved recognition rates up to 99.25% on a dataset of 2,000 handwritten digits. Statistical and structural features are extracted to represent characters for classification using deformable templates.
Review and comparison of tasks scheduling in cloud computingijfcstjournal
Recently, there has been a dramatic increase in the popularity of cloud computing systems that rent
computing resources on-demand, bill on a pay-as-you-go basis, and multiplex many users on the same
physical infrastructure. It is a virtual pool of resources which are provided to users via Internet. It gives
users virtually unlimited pay-per-use computing resources without the burden of managing the underlying
infrastructure. One of the goals is to use the resources efficiently and gain maximum profit. Scheduling is a
critical problem in Cloud computing, because a cloud provider has to serve many users in Cloud
computing system. So scheduling is the major issue in establishing Cloud computing systems. The
scheduling algorithms should order the jobs in a way where balance between improving the performance
and quality of service and at the same time maintaining the efficiency and fairness among the jobs. This
paper introduces and explores some of the methods provided for in cloud computing has been scheduled.
Finally the waiting time and time to implement some of the proposed algorithm is evaluated
1. The document presents a methodology for recognizing isolated handwritten Devanagari numerals using structural and statistical features.
2. Key features extracted include whether the numeral has openings on the left, right, above or below, and the number of horizontal and vertical crossings.
3. The methodology achieves an average accuracy of 96.8% on a dataset of 500 numeral images collected from various individuals. Accuracy is highest for numerals 0, 6, 8 and 10 at 100%, while some similar numerals like 3 and 2 see more errors.
GUI based handwritten digit recognition using CNNAbhishek Tiwari
This project is to create a model which can recognize the digits as well as also to create GUI which is user friendly i.e. user can draw the digit on it and will get appropriate output.
Artificial Neural Network For Recognition Of Handwritten Devanagari CharacterIOSR Journals
1) The document discusses recognizing handwritten Devanagari characters using artificial neural networks and zone-based feature extraction.
2) It proposes extracting features from images by dividing them into zones and calculating average pixel distances to the image and zone centroids.
3) This zone-based feature vector is then input to a feedforward neural network for character recognition.
An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...paperpublications3
Abstract: This paper presents a new approach using Hidden Markov Model as classifier and Singular Values Decomposition (SVD) coefficients as features for iris recognition. As iris is a complex multi-dimensional structure and needs good computing techniques for recognition and it is an integral part of biometrics. Features extracted from a iris are processed and compared with similar irises which exist in database. The recognition of human irises is carried out by comparing characteristics of the iris to those of known individuals. Here seven state Hidden Markov Model (HMM)-based iris recognition system is proposed .A small number of quantized Singular Value Decomposition (SVD) coefficients as features describing blocks of iris images. SVD is a method for transforming correlated variables into a set of uncorrelated ones that better expose the various relationships among the original data item. This makes the system very fast. The proposed approach has been examined on CASIA database. The results show that the proposed method is the fastest one, having good accuracy.
A novel embedded hybrid thinning algorithm forprjpublications
The document proposes a hybrid thinning algorithm that combines the Stentiford and Zhang-Suen thinning algorithms. It compares the hybrid algorithm to the original Stentiford and Zhang-Suen algorithms on an input image. The hybrid algorithm more accurately thins the image to a single pixel width but does not improve time complexity compared to the original algorithms. The hybrid approach uses four templates across two sub-iterations to identify and remove pixels based on connectivity values until no more can be removed. Experimental results show the hybrid algorithm more effectively increases image contrast than the original thinning algorithms.
This document provides an exploratory review of soft computing techniques for image segmentation. It discusses various segmentation techniques including discontinuity-based techniques like point, line and edge detection using spatial filtering. Thresholding techniques like global, adaptive and multi-level thresholding are also covered. Region-based techniques such as region growing, region splitting/merging and morphological watersheds are summarized. The document concludes that future work can focus on developing genetic segmentation filters using a genetic algorithm approach for medical image segmentation.
This document discusses automatic image annotation using weakly supervised graph propagation. It begins with an abstract describing the method, which assigns annotated labels to semantically derived image regions using training images, pre-assigned labels, and input images. Graph construction considers consistency and incongruity relationships between image patches. Label propagation aims to propagate labels from images to patches using these relationships. Evaluation shows the proposed method achieves higher accuracy than baselines on several datasets by leveraging contextual information between image regions.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
A Pattern Classification Based approach for Blur Classificationijeei-iaes
Blur type identification is one of the most crucial step of image restoration. In case of blind restoration of such images, it is generally assumed that the blur type is known prior to restoration of such images. However, it is not practical in real applications. So, blur type identification is extremely desirable before application of blind restoration technique to restore a blurred image. An approach to categorize blur in three classes namely motion, defocus, and combined blur is presented in this paper. Curvelet transform based energy features are utilized as features of blur patterns and a neural network is designed for classification. The simulation results show preciseness of proposed approach.
This document discusses and compares two partition-based clustering algorithms, K-means and K-medoids, for image segmentation. It provides an overview of image segmentation and clustering, describes the K-means and K-medoids algorithms, and analyzes their pros and cons. Key differences between the two algorithms are their complexity, sensitivity to noise, flexibility, and whether they use the mean or medoid as the cluster representative. The document concludes that both algorithms can effectively segment images but K-medoids is more robust to noise while K-means has lower computational cost.
This document summarizes a research paper on color image segmentation using k-means clustering. It discusses how k-means clustering can be used to group color image pixels into a set number of classes without using training data. The clustering groups similar color pixels to obtain segmentation. This avoids calculating features for every pixel and provides efficient segmentation based on color similarity. The document outlines the k-means clustering process used and how it segments an image into distinct colored regions to extract important objects.
This document presents a blur classification approach using a Convolution Neural Network (CNN). It discusses types of image degradation including blur, different blur models, and prior work on blur classification using features and neural networks. The proposed method uses a CNN to classify images into four blur categories (motion, defocus, box, and Gaussian blur) based on the images' frequency spectra. The method is evaluated on a dataset with over 2800 synthetically blurred images from 24 people performing 10 gestures. The CNN achieves an average accuracy of 97% for blur classification, outperforming alternatives using multilayer perceptrons or handcrafted features.
RECOGNITION OF HANDWRITTEN MEITEI MAYEK SCRIPT BASED ON TEXTURE FEATURE kevig
Recognition of Manipuri Script called Meitei Mayek is still in the infant stage due to its complex structure. In this paper, an attempt has been made to develop an offline Meitei Mayek handwritten character recognition model by exploiting the texture feature, Local Binary Pattern (LBP). The system has been developed and evaluated on a large dataset consisting of 3,780 characters which are collected from different people of varying age group. The highest recognition rate achieved by the proposed method is 93.33% using Support Vector Machine (SVM). So, the contribution of this paper is bi-fold: firstly, a collection of a large handwritten corpus of Meitei Mayek Script and secondly developing character recognition model on the collected dataset.
This document discusses text detection and character recognition from images. It begins with an introduction and then discusses the aims, objectives, motivation and problem statement. It reviews relevant literature on segmentation and recognition techniques. The document then describes the methodology used, including preprocessing, segmentation using vertical projections and connected components, and recognition using pixel counting, projections, template matching, Fourier descriptors and heuristic filters. It presents results from four experiments comparing different segmentation and recognition methods. The discussion analyzes results and limitations. The conclusion finds that segmentation works best with connected components while recognition works best with template matching, Fourier descriptors and heuristic filters.
Image classification is perhaps the most important part of digital image analysis. In this paper, we compare the most widely used model CNN Convolutional Neural Network , and MLP Multilayer Perceptron . We aim to show how both models differ and how both models approach towards the final goal, which is image classification. Souvik Banerjee | Dr. A Rengarajan "Hand-Written Digit Classification" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-4 , June 2021, URL: https://www.ijtsrd.compapers/ijtsrd42444.pdf Paper URL: https://www.ijtsrd.comcomputer-science/artificial-intelligence/42444/handwritten-digit-classification/souvik-banerjee
A Secure Color Image Steganography in Transform Domain ijcisjournal
Steganography is the art and science of covert communication. The secret information can be concealed in content such as image, audio, or video. This paper provides a novel image steganography technique to hide both image and key in color cover image using Discrete Wavelet Transform (DWT) and Integer Wavelet Transform (IWT). There is no visual difference between the stego image and the cover image. The extracted image is also similar to the secret image. This is proved by the high PSNR (Peak Signal to Noise Ratio), value for both stego and extracted secret image. The results are compared with the results of similar techniques and it is found that the proposed technique is simple and gives better PSNR values than others.
This document discusses object recognition by computers. It notes that while object recognition is easy for humans, it is difficult for computers because they cannot rely on appearance alone. Key challenges for computers include variations in scale, shape, occlusion, lighting and background clutter. The document then discusses techniques used for object recognition, including feature detection methods like SIFT and SURF that extract keypoints, descriptors that describe regions around keypoints, and feature matching to identify corresponding regions between images. It also covers bag-of-words models, visual vocabularies and inverted indexing to allow large scale image retrieval. Finally, it lists applications of object recognition like digital watermarking, face detection and robot navigation.
IRJET- Image Captioning using Multimodal EmbeddingIRJET Journal
This document proposes a novel methodology to generate a single story or caption from multiple images that share similar context. It combines existing image captioning and natural language processing models. Specifically, it uses a Convolutional Neural Network to extract visual semantics from images and generate captions. It then represents the captions as vectors using Skip Thought vectors or TF-IDF values. These vectors are combined into a matrix and used to generate a new story/caption that shares the context of the input images. The results show that the Skip Thought vector approach achieves better performance based on RMSE and MAE error metrics. The model could potentially be applied to applications like medical diagnosis, crime investigations, and lecture note generation.
Text Extraction of Colour Images using Mathematical Morphology & HAAR TransformIOSR Journals
This document summarizes a research paper that proposes a method to extract text from color images using the Haar discrete wavelet transform (DWT) and mathematical morphology. It begins with an introduction to text extraction challenges and prior approaches. The proposed method detects text edges using the Haar DWT and removes non-text edges with thresholding. Morphological dilation is then used to connect isolated candidate text edges. Mathematical morphology and templates are also used to extract characters from images. The simulation was carried out using MATLAB for image processing.
11.development of a writer independent online handwritten character recogniti...Alexander Decker
This document describes the development of an online handwritten character recognition system using a modified hybrid neural network model. It developed a hybrid feature extraction technique that combines stroke information, contour pixels, and zoning of characters to create feature vectors. A hybrid neural network model combining modified counterpropagation and optical backpropagation networks was also developed. Experiments using 6,200 character samples from 50 subjects achieved a 99% recognition rate with an average recognition time of 2 milliseconds when testing samples from new subjects.
Optical Character Recognition from Text ImageEditor IJCATR
Optical Character Recognition (OCR) is a system that provides a full alphanumeric recognition of printed or handwritten
characters by simply scanning the text image. OCR system interprets the printed or handwritten characters image and converts it into
corresponding editable text document. The text image is divided into regions by isolating each line, then individual characters with
spaces. After character extraction, the texture and topological features like corner points, features of different regions, ratio of
character area and convex area of all characters of text image are calculated. Previously features of each uppercase and lowercase
letter, digit, and symbols are stored as a template. Based on the texture and topological features, the system recognizes the exact
character using feature matching between the extracted character and the template of all characters as a measure of similarity.
1. The document presents a methodology for recognizing isolated handwritten Devanagari numerals using structural and statistical features.
2. Key features extracted include whether the numeral has openings on the left, right, above or below, and the number of horizontal and vertical crossings.
3. The methodology achieves an average accuracy of 96.8% on a dataset of 500 numeral images collected from various individuals. Accuracy is highest for numerals 0, 6, 8 and 10 at 100%, while some similar numerals like 3 and 2 see more errors.
GUI based handwritten digit recognition using CNNAbhishek Tiwari
This project is to create a model which can recognize the digits as well as also to create GUI which is user friendly i.e. user can draw the digit on it and will get appropriate output.
Artificial Neural Network For Recognition Of Handwritten Devanagari CharacterIOSR Journals
1) The document discusses recognizing handwritten Devanagari characters using artificial neural networks and zone-based feature extraction.
2) It proposes extracting features from images by dividing them into zones and calculating average pixel distances to the image and zone centroids.
3) This zone-based feature vector is then input to a feedforward neural network for character recognition.
An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...paperpublications3
Abstract: This paper presents a new approach using Hidden Markov Model as classifier and Singular Values Decomposition (SVD) coefficients as features for iris recognition. As iris is a complex multi-dimensional structure and needs good computing techniques for recognition and it is an integral part of biometrics. Features extracted from a iris are processed and compared with similar irises which exist in database. The recognition of human irises is carried out by comparing characteristics of the iris to those of known individuals. Here seven state Hidden Markov Model (HMM)-based iris recognition system is proposed .A small number of quantized Singular Value Decomposition (SVD) coefficients as features describing blocks of iris images. SVD is a method for transforming correlated variables into a set of uncorrelated ones that better expose the various relationships among the original data item. This makes the system very fast. The proposed approach has been examined on CASIA database. The results show that the proposed method is the fastest one, having good accuracy.
A novel embedded hybrid thinning algorithm forprjpublications
The document proposes a hybrid thinning algorithm that combines the Stentiford and Zhang-Suen thinning algorithms. It compares the hybrid algorithm to the original Stentiford and Zhang-Suen algorithms on an input image. The hybrid algorithm more accurately thins the image to a single pixel width but does not improve time complexity compared to the original algorithms. The hybrid approach uses four templates across two sub-iterations to identify and remove pixels based on connectivity values until no more can be removed. Experimental results show the hybrid algorithm more effectively increases image contrast than the original thinning algorithms.
This document provides an exploratory review of soft computing techniques for image segmentation. It discusses various segmentation techniques including discontinuity-based techniques like point, line and edge detection using spatial filtering. Thresholding techniques like global, adaptive and multi-level thresholding are also covered. Region-based techniques such as region growing, region splitting/merging and morphological watersheds are summarized. The document concludes that future work can focus on developing genetic segmentation filters using a genetic algorithm approach for medical image segmentation.
This document discusses automatic image annotation using weakly supervised graph propagation. It begins with an abstract describing the method, which assigns annotated labels to semantically derived image regions using training images, pre-assigned labels, and input images. Graph construction considers consistency and incongruity relationships between image patches. Label propagation aims to propagate labels from images to patches using these relationships. Evaluation shows the proposed method achieves higher accuracy than baselines on several datasets by leveraging contextual information between image regions.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
A Pattern Classification Based approach for Blur Classificationijeei-iaes
Blur type identification is one of the most crucial step of image restoration. In case of blind restoration of such images, it is generally assumed that the blur type is known prior to restoration of such images. However, it is not practical in real applications. So, blur type identification is extremely desirable before application of blind restoration technique to restore a blurred image. An approach to categorize blur in three classes namely motion, defocus, and combined blur is presented in this paper. Curvelet transform based energy features are utilized as features of blur patterns and a neural network is designed for classification. The simulation results show preciseness of proposed approach.
This document discusses and compares two partition-based clustering algorithms, K-means and K-medoids, for image segmentation. It provides an overview of image segmentation and clustering, describes the K-means and K-medoids algorithms, and analyzes their pros and cons. Key differences between the two algorithms are their complexity, sensitivity to noise, flexibility, and whether they use the mean or medoid as the cluster representative. The document concludes that both algorithms can effectively segment images but K-medoids is more robust to noise while K-means has lower computational cost.
This document summarizes a research paper on color image segmentation using k-means clustering. It discusses how k-means clustering can be used to group color image pixels into a set number of classes without using training data. The clustering groups similar color pixels to obtain segmentation. This avoids calculating features for every pixel and provides efficient segmentation based on color similarity. The document outlines the k-means clustering process used and how it segments an image into distinct colored regions to extract important objects.
This document presents a blur classification approach using a Convolution Neural Network (CNN). It discusses types of image degradation including blur, different blur models, and prior work on blur classification using features and neural networks. The proposed method uses a CNN to classify images into four blur categories (motion, defocus, box, and Gaussian blur) based on the images' frequency spectra. The method is evaluated on a dataset with over 2800 synthetically blurred images from 24 people performing 10 gestures. The CNN achieves an average accuracy of 97% for blur classification, outperforming alternatives using multilayer perceptrons or handcrafted features.
RECOGNITION OF HANDWRITTEN MEITEI MAYEK SCRIPT BASED ON TEXTURE FEATURE kevig
Recognition of Manipuri Script called Meitei Mayek is still in the infant stage due to its complex structure. In this paper, an attempt has been made to develop an offline Meitei Mayek handwritten character recognition model by exploiting the texture feature, Local Binary Pattern (LBP). The system has been developed and evaluated on a large dataset consisting of 3,780 characters which are collected from different people of varying age group. The highest recognition rate achieved by the proposed method is 93.33% using Support Vector Machine (SVM). So, the contribution of this paper is bi-fold: firstly, a collection of a large handwritten corpus of Meitei Mayek Script and secondly developing character recognition model on the collected dataset.
This document discusses text detection and character recognition from images. It begins with an introduction and then discusses the aims, objectives, motivation and problem statement. It reviews relevant literature on segmentation and recognition techniques. The document then describes the methodology used, including preprocessing, segmentation using vertical projections and connected components, and recognition using pixel counting, projections, template matching, Fourier descriptors and heuristic filters. It presents results from four experiments comparing different segmentation and recognition methods. The discussion analyzes results and limitations. The conclusion finds that segmentation works best with connected components while recognition works best with template matching, Fourier descriptors and heuristic filters.
Image classification is perhaps the most important part of digital image analysis. In this paper, we compare the most widely used model CNN Convolutional Neural Network , and MLP Multilayer Perceptron . We aim to show how both models differ and how both models approach towards the final goal, which is image classification. Souvik Banerjee | Dr. A Rengarajan "Hand-Written Digit Classification" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-4 , June 2021, URL: https://www.ijtsrd.compapers/ijtsrd42444.pdf Paper URL: https://www.ijtsrd.comcomputer-science/artificial-intelligence/42444/handwritten-digit-classification/souvik-banerjee
A Secure Color Image Steganography in Transform Domain ijcisjournal
Steganography is the art and science of covert communication. The secret information can be concealed in content such as image, audio, or video. This paper provides a novel image steganography technique to hide both image and key in color cover image using Discrete Wavelet Transform (DWT) and Integer Wavelet Transform (IWT). There is no visual difference between the stego image and the cover image. The extracted image is also similar to the secret image. This is proved by the high PSNR (Peak Signal to Noise Ratio), value for both stego and extracted secret image. The results are compared with the results of similar techniques and it is found that the proposed technique is simple and gives better PSNR values than others.
This document discusses object recognition by computers. It notes that while object recognition is easy for humans, it is difficult for computers because they cannot rely on appearance alone. Key challenges for computers include variations in scale, shape, occlusion, lighting and background clutter. The document then discusses techniques used for object recognition, including feature detection methods like SIFT and SURF that extract keypoints, descriptors that describe regions around keypoints, and feature matching to identify corresponding regions between images. It also covers bag-of-words models, visual vocabularies and inverted indexing to allow large scale image retrieval. Finally, it lists applications of object recognition like digital watermarking, face detection and robot navigation.
IRJET- Image Captioning using Multimodal EmbeddingIRJET Journal
This document proposes a novel methodology to generate a single story or caption from multiple images that share similar context. It combines existing image captioning and natural language processing models. Specifically, it uses a Convolutional Neural Network to extract visual semantics from images and generate captions. It then represents the captions as vectors using Skip Thought vectors or TF-IDF values. These vectors are combined into a matrix and used to generate a new story/caption that shares the context of the input images. The results show that the Skip Thought vector approach achieves better performance based on RMSE and MAE error metrics. The model could potentially be applied to applications like medical diagnosis, crime investigations, and lecture note generation.
Text Extraction of Colour Images using Mathematical Morphology & HAAR TransformIOSR Journals
This document summarizes a research paper that proposes a method to extract text from color images using the Haar discrete wavelet transform (DWT) and mathematical morphology. It begins with an introduction to text extraction challenges and prior approaches. The proposed method detects text edges using the Haar DWT and removes non-text edges with thresholding. Morphological dilation is then used to connect isolated candidate text edges. Mathematical morphology and templates are also used to extract characters from images. The simulation was carried out using MATLAB for image processing.
11.development of a writer independent online handwritten character recogniti...Alexander Decker
This document describes the development of an online handwritten character recognition system using a modified hybrid neural network model. It developed a hybrid feature extraction technique that combines stroke information, contour pixels, and zoning of characters to create feature vectors. A hybrid neural network model combining modified counterpropagation and optical backpropagation networks was also developed. Experiments using 6,200 character samples from 50 subjects achieved a 99% recognition rate with an average recognition time of 2 milliseconds when testing samples from new subjects.
Optical Character Recognition from Text ImageEditor IJCATR
Optical Character Recognition (OCR) is a system that provides a full alphanumeric recognition of printed or handwritten
characters by simply scanning the text image. OCR system interprets the printed or handwritten characters image and converts it into
corresponding editable text document. The text image is divided into regions by isolating each line, then individual characters with
spaces. After character extraction, the texture and topological features like corner points, features of different regions, ratio of
character area and convex area of all characters of text image are calculated. Previously features of each uppercase and lowercase
letter, digit, and symbols are stored as a template. Based on the texture and topological features, the system recognizes the exact
character using feature matching between the extracted character and the template of all characters as a measure of similarity.
10 Insightful Quotes On Designing A Better Customer ExperienceYuan Wang
In an ever-changing landscape of one digital disruption after another, companies and organisations are looking for new ways to understand their target markets and engage them better. Increasingly they invest in user experience (UX) and customer experience design (CX) capabilities by working with a specialist UX agency or developing their own UX lab. Some UX practitioners are touting leaner and faster ways of developing customer-centric products and services, via methodologies such as guerilla research, rapid prototyping and Agile UX. Others seek innovation and fulfilment by spending more time in research, being more inclusive, and designing for social goods.
Experience is more than just an interface. It is a relationship, as well as a series of touch points between your brand and your customer. Here are our top 10 highlights and takeaways from the recent UX Australia conference to help you transform your customer experience design.
For full article, continue reading at https://yump.com.au/10-ways-supercharge-customer-experience-design/
http://inarocket.com
Learn BEM fundamentals as fast as possible. What is BEM (Block, element, modifier), BEM syntax, how it works with a real example, etc.
How to Build a Dynamic Social Media PlanPost Planner
Stop guessing and wasting your time on networks and strategies that don’t work!
Join Rebekah Radice and Katie Lance to learn how to optimize your social networks, the best kept secrets for hot content, top time management tools, and much more!
Watch the replay here: bit.ly/socialmedia-plan
The document discusses how personalization and dynamic content are becoming increasingly important on websites. It notes that 52% of marketers see content personalization as critical and 75% of consumers like it when brands personalize their content. However, personalization can create issues for search engine optimization as dynamic URLs and content are more difficult for search engines to index than static pages. The document provides tips for SEOs to help address these personalization and SEO challenges, such as using static URLs when possible and submitting accurate sitemaps.
This document discusses a method for handwritten character recognition using a K-nearest neighbors (K-NN) classification algorithm. It begins by introducing the problem of handwritten character recognition and the challenges involved. It then describes the main steps of the proposed method: preprocessing the image data, extracting features, and classifying characters using K-NN. The document tests the method on the MNIST dataset of handwritten digits, achieving an accuracy of 97.67%. It concludes that the method is able to accurately recognize handwritten characters independently of size, font, or writer style.
SEGMENTATION AND RECOGNITION OF HANDWRITTEN DIGIT NUMERAL STRING USING A MULT...ijfcstjournal
In this paper, the use of Multi-Layer Perceptron (MLP) Neural Network model is proposed for recognizing
unconstrained offline handwritten Numeral strings. The Numeral strings are segmented and isolated
numerals are obtained using a connected component labeling (CCL) algorithm approach. The structural
part of the models has been modeled using a Multilayer Perceptron Neural Network. This paper also
presents a new technique to remove slope and slant from handwritten numeral string and to normalize the
size of text images and classify with supervised learning methods. Experimental results on a database of
102 numeral string patterns written by 3 different people show that a recognition rate of 99.7% is obtained
on independent digits contained in the numeral string of digits includes both the skewed and slant data.
1) The document discusses recognizing handwritten Devanagari characters using an artificial neural network approach.
2) It proposes extracting features from images by dividing them into zones and calculating average pixel distances to the image and zone centroids.
3) This zone-based feature extraction method is used to create a feature vector for each character image that is then classified by a feedforward neural network.
Optically processed Kannada script realization with Siamese neural network modelIAESIJAI
Optical character recognition (OCR) is a technology that allows computers to recognize and extract text from images or scanned documents. It is commonly used to convert printed or handwritten text into machine-readable format. This Study presents an OCR system on Kannada Characters based on siamese neural network (SNN). Here the SNN, a Deep neural network which comprises of two identical convolutional neural network (CNN) compare the script and ranks based on the dissimilarity. When lesser dissimilarity score is identified, prediction is done as character match. In this work the authors use 5 classes of Kannada characters which were initially preprocessed using grey scaling and convert it to pgm format. This is directly input into the Deep convolutional network which is learnt from matching and non-matching image between the CNN with contrastive loss function in Siamese architecture. The Proposed OCR system uses very less time and gives more accurate results as compared to the regular CNN. The model can become a powerful tool for identification, particularly in situations where there is a high degree of variation in writing styles or limited training data is available.
This document summarizes research on recognizing handwritten characters in the Odia language. It discusses two main approaches to Odia character recognition: template matching and feature extraction. The document also reviews several papers on Odia handwritten character recognition, describing the different techniques used, such as neural networks, genetic algorithms, and rule-based methods. Overall, the document surveys existing work on developing systems for Odia optical character recognition (OCR) and handwritten character recognition.
Devanagari Digit and Character Recognition Using Convolutional Neural NetworkIRJET Journal
This document describes a system for recognizing handwritten Devanagari digits and characters using a convolutional neural network (CNN). The system is designed to overcome challenges from variations in handwriting styles. It involves preprocessing the dataset, extracting features, training a CNN model on training images, and using the trained model to classify testing and real-time input images and output the recognized character or digit. An experiment using a Kaggle dataset of 92,000 Devanagari character and digit images achieved recognition of user-drawn input on an interface using the trained CNN model.
Faster Training Algorithms in Neural Network Based Approach For Handwritten T...CSCJournals
Handwritten text and character recognition is a challenging task compared to recognition of handwritten numeral and computer printed text due to its large variety in nature. As practical pattern recognition problems uses bulk data and there is a one step self sufficient deterministic theory to resolve recognition problems by calculating inverse of Hessian Matrix and multiplication the inverse matrix it with first order local gradient vector. But in practical cases when neural network is large the inversing operation of the Hessian Matrix is not manageable and another condition must be satisfied the Hessian Matrix must be positive definite which may not be satishfied. In these cases some repetitive recursive models are taken. In several research work in past decade it was experienced that Neural Network based approach provides most reliable performance in handwritten character and text recognition but recognition performance depends upon some important factors like no of training samples, reliable features and no of features per character, training time, variety of handwriting etc. Important features from different types of handwriting are collected and are fed to the neural network for training. It is true that more no of features increases test efficiency but it takes longer time to converge the error curve. To reduce this training time effectively proper train algorithm should be chosen so that the system provides best train and test efficiency in least possible time that is to provide the system fastest intelligence. We have used several second order conjugate gradient algorithms for training of neural network. We have found that Scaled Conjugate Gradient Algorithm , a second order training algorithm as the fastest for training of neural network for our application. Training using SCG takes minimum time with excellent test efficiency. A scanned handwritten text is taken as input and character level segmentation is done. Some important and reliable features from each character are extracted and used as input to a neural network for training. When the error level reaches into a satisfactory level (10 -12 ) weights are accepted for testing a test script. Finally a lexicon matching algorithm solves the minor misclassification problems.
11.secure compressed image transmission using self organizing feature mapsAlexander Decker
This document summarizes a research paper that proposes a method for secure compressed image transmission using self-organizing feature maps. The method involves compressing images using SOFM-based vector quantization, entropy coding the results, and encrypting the compressed data using a scrambler before transmission. Simulation results show the method achieves a compression ratio of up to 38:1 while providing security, outperforming JPEG compression by up to 1 dB. The paper presents the technical details and evaluation of the proposed secure image transmission system.
Text detection and recognition in scene images or natural images has applications in computer
vision systems like registration number plate detection, automatic traffic sign detection, image retrieval
and help for visually impaired people. Scene text, however, has complicated background, blur image,
partly occluded text, variations in font-styles, image noise and ranging illumination. Hence scene text
recognition could be a difficult computer vision problem. In this paper connected component method
is used to extract the text from background. In this work, horizontal and vertical projection profiles,
geometric properties of text, image binirization and gap filling method are used to extract the text from
scene images. Then histogram based threshold is applied to separate text background of the images.
Finally text is extracted from images.
International Journal of Research in Engineering and Science is an open access peer-reviewed international forum for scientists involved in research to publish quality and refereed papers. Papers reporting original research or experimentally proved review work are welcome. Papers for publication are selected through peer review to ensure originality, relevance, and readability.
OCR for Gujarati Numeral using Neural Networkijsrd.com
This papers functions within to reduce individuality popularity (OCR) program for hand-written Gujarati research. One can find so much of work for Indian own native different languages like Hindi, Gujarati, Tamil, Bengali, Malayalam, Gurumukhi etc., but Gujarati is a vocabulary for which hardly any work is traceable especially for hand-written individuals. Here in this work a nerve program is provided for Gujarati hand-written research popularity. This paper deals with an optical character recognition (OCR) system for handwritten Gujarati numbers. A several break up food ahead nerve program is suggested for variation of research. The functions of Gujarati research are abstracted by four different details of research. Reduction and skew- changes are also done for preprocessing of hand-written research before their variation. This work has purchased approximately 81% of performance for Gujarati handwritten numerals.
AUTOMATIC TRAINING DATA SYNTHESIS FOR HANDWRITING RECOGNITION USING THE STRUC...ijaia
The paper presents a novel technique called “Structural Crossing-Over” to synthesize qualified data for training machine learning-based handwriting recognition. The proposed technique can provide a greater variety of patterns of training data than the existing approaches such as elastic distortion and tangentbased affine transformation. A couple of training characters are chosen, then they are analyzed by their similar and different structures, and finally are crossed over to generate the new characters. The experiments are set to compare the performances of tangent-based affine transformation and the proposed approach in terms of the variety of generated characters and percent of recognition errors. The standard
MNIST corpus including 60,000 training characters and 10,000 test characters is employed in the
experiments. The proposed technique uses 1,000 characters to synthesize 60,000 characters, and then uses
these data to train and test the benchmark handwriting recognition system that exploits Histogram of
Gradient: HOG as features and Support Vector Machine: SVM as recognizer. The experimental result
yields 8.06% of errors. It significantly outperforms the tangent-based affine transformation and the
original MNIST training data, which are 11.74% and 16.55%, respectively.
Performance analysis of chain code descriptor for hand shape classificationijcga
Feature Extraction is an important task for any Image processing application. The visual properties of any image are its shape, texture and colour. Out of these shape description plays important role in any image classification. The shape description method classified into two types, contour base and region based. The contour base method concentrated on the shape boundary line and the region based method considers whole area. In this paper, contour based, the chain code description method was experimented for different hand shape.
The chain code descriptor of various hand shapes was calculated and tested with different classifier such as k-nearest- neighbour (k-NN), Support vector machine (SVM) and Naive Bayes. Principal component analysis (PCA) was applied after the chain code description. The performance of SVM was found better than k-NN and Naive Bayes with recognition rate 93%.
A SURVEY ON DEEP LEARNING METHOD USED FOR CHARACTER RECOGNITIONIJCIRAS Journal
The field of Artificial Intelligence is very fashionable today, especially neural networks that work well in various areas such as speech recognition and natural language processing. This Research Article briefly describes how deep learning models work and what different techniques are used in text recognition. It also describes the great progress that has been made in the field of medicine, the analysis of forensic documents, the recognition of license plates, banking, health and the legal industry. The recognition of handwritten characters is one of the research areas in the field of artificial intelligence. The individual character recognition has a higher recognition accuracy than the complete word recognition. The new method for categorizing Freeman strings is presented using four connectivity events and eight connectivity events with a deep learning approach.
offline character recognition for handwritten gujarati textBhumika Patel
This document summarizes a presentation on optical character recognition of Gujarati characters using convolutional neural networks. It outlines collecting a dataset of 1360 images each of 34 Gujarati characters written by different people. The proposed approach involves preprocessing images, training a CNN model, and calculating accuracy. Initial results correctly recognized some characters but had difficulty with connected characters. Future work includes recognizing remaining characters and vowels, collecting more data, and exploring different CNN configurations to improve accuracy.
This document discusses methods for one-shot learning using siamese neural networks. It provides an overview of several key papers in this area, including using siamese networks for signature verification (1993) and one-shot image recognition (2015), and introducing matching networks for one-shot learning (2016). Matching networks incorporate an attention mechanism into a neural network to rapidly learn from small datasets by matching training and test conditions. The document also reviews experiments demonstrating one-shot and few-shot learning on datasets like Omniglot using these siamese and matching network approaches.
Hand Written Character Recognition Using Neural Networks Chiranjeevi Adi
This document discusses a project to develop a handwritten character recognition system using a neural network. It will take handwritten English characters as input and recognize the patterns using a trained neural network. The system aims to recognize individual characters as well as classify them into groups. It will first preprocess, segment, extract features from, and then classify the input characters using the neural network. The document reviews several existing approaches to handwritten character recognition and the use of gradient and edge-based feature extraction with neural networks. It defines the objectives and methods for the proposed system, which will involve preprocessing, segmentation, feature extraction, and classification/recognition steps. Finally, it outlines the hardware and software requirements to implement the system as a MATLAB application.
Feature Extraction and Feature Selection using Textual Analysisvivatechijri
After pre-processing the images in character recognition systems, the images are segmented based on
certain characteristics known as “features”. The feature space identified for character recognition is however
ranging across a huge dimensionality. To solve this problem of dimensionality, the feature selection and feature
extraction methods are used. Hereby in this paper, we are going to discuss, the different techniques for feature
extraction and feature selection and how these techniques are used to reduce the dimensionality of feature space
to improve the performance of text categorization.
Nowadays character recognition has gained lot of attention in the field of pattern recognition due to its application in various fields. It is one of the most successful applications of automatic pattern recognition. Research in OCR is popular for its application potential in banks, post offices, office automation etc. HCR is useful in cheque processing in banks; almost all kind of form processing systems, handwritten postal address resolution and many more. This paper presents a simple and efficient approach for the implementation of OCR and translation of scanned images of printed text into machine-encoded text. It makes use of different image analysis phases followed by image detection via pre-processing and post-processing. This paper also describes scanning the entire document (same as the segmentation in our case) and recognizing individual characters from image irrespective of their position, size and various font styles and it deals with recognition of the symbols from English language, which is internationally accepted.
IRJET- Intelligent Character Recognition of Handwritten Characters using ...IRJET Journal
This document describes a system for intelligent character recognition of handwritten characters using an artificial neural network. The system takes images of handwritten characters as input and trains a neural network using stochastic gradient descent and backpropagation to classify the characters. After training, the neural network can recognize characters in new input images, even when there is noise in the images. The system aims to improve on existing optical character recognition techniques by using an artificial neural network approach that can recognize a variety of handwriting styles and fonts.
Unlock your organization’s full potential with the 2025 Digital Adoption Blueprint. Discover proven strategies to streamline software onboarding, boost productivity, and drive enterprise-wide digital transformation.
New Ways to Reduce Database Costs with ScyllaDBScyllaDB
How ScyllaDB’s latest capabilities can reduce your infrastructure costs
ScyllaDB has been obsessed with price-performance from day 1. Our core database is architected with low-level engineering optimizations that squeeze every ounce of power from the underlying infrastructure. And we just completed a multi-year effort to introduce a set of new capabilities for additional savings.
Join this webinar to learn about these new capabilities: the underlying challenges we wanted to address, the workloads that will benefit most from each, and how to get started. We’ll cover ways to:
- Avoid overprovisioning with “just-in-time” scaling
- Safely operate at up to ~90% storage utilization
- Cut network costs with new compression strategies and file-based streaming
We’ll also highlight a “hidden gem” capability that lets you safely balance multiple workloads in a single cluster. To conclude, we will share the efficiency-focused capabilities on our short-term and long-term roadmaps.
Supercharge Your AI Development with Local LLMsFrancesco Corti
In today's AI development landscape, developers face significant challenges when building applications that leverage powerful large language models (LLMs) through SaaS platforms like ChatGPT, Gemini, and others. While these services offer impressive capabilities, they come with substantial costs that can quickly escalate especially during the development lifecycle. Additionally, the inherent latency of web-based APIs creates frustrating bottlenecks during the critical testing and iteration phases of development, slowing down innovation and frustrating developers.
This talk will introduce the transformative approach of integrating local LLMs directly into their development environments. By bringing these models closer to where the code lives, developers can dramatically accelerate development lifecycles while maintaining complete control over model selection and configuration. This methodology effectively reduces costs to zero by eliminating dependency on pay-per-use SaaS services, while opening new possibilities for comprehensive integration testing, rapid prototyping, and specialized use cases.
Adtran’s new Ensemble Cloudlet vRouter solution gives service providers a smarter way to replace aging edge routers. With virtual routing, cloud-hosted management and optional design services, the platform makes it easy to deliver high-performance Layer 3 services at lower cost. Discover how this turnkey, subscription-based solution accelerates deployment, supports hosted VNFs and helps boost enterprise ARPU.
UiPath Community Berlin: Studio Tips & Tricks and UiPath InsightsUiPathCommunity
Join the UiPath Community Berlin (Virtual) meetup on May 27 to discover handy Studio Tips & Tricks and get introduced to UiPath Insights. Learn how to boost your development workflow, improve efficiency, and gain visibility into your automation performance.
📕 Agenda:
- Welcome & Introductions
- UiPath Studio Tips & Tricks for Efficient Development
- Best Practices for Workflow Design
- Introduction to UiPath Insights
- Creating Dashboards & Tracking KPIs (Demo)
- Q&A and Open Discussion
Perfect for developers, analysts, and automation enthusiasts!
This session streamed live on May 27, 18:00 CET.
Check out all our upcoming UiPath Community sessions at:
👉 https://community.uipath.com/events/
Join our UiPath Community Berlin chapter:
👉 https://community.uipath.com/berlin/
Create Your First AI Agent with UiPath Agent BuilderDianaGray10
Join us for an exciting virtual event where you'll learn how to create your first AI Agent using UiPath Agent Builder. This session will cover everything you need to know about what an agent is and how easy it is to create one using the powerful AI-driven UiPath platform. You'll also discover the steps to successfully publish your AI agent. This is a wonderful opportunity for beginners and enthusiasts to gain hands-on insights and kickstart their journey in AI-powered automation.
Measuring Microsoft 365 Copilot and Gen AI SuccessNikki Chapple
Session | Measuring Microsoft 365 Copilot and Gen AI Success with Viva Insights and Purview
Presenter | Nikki Chapple 2 x MVP and Principal Cloud Architect at CloudWay
Event | European Collaboration Conference 2025
Format | In person Germany
Date | 28 May 2025
📊 Measuring Copilot and Gen AI Success with Viva Insights and Purview
Presented by Nikki Chapple – Microsoft 365 MVP & Principal Cloud Architect, CloudWay
How do you measure the success—and manage the risks—of Microsoft 365 Copilot and Generative AI (Gen AI)? In this ECS 2025 session, Microsoft MVP and Principal Cloud Architect Nikki Chapple explores how to go beyond basic usage metrics to gain full-spectrum visibility into AI adoption, business impact, user sentiment, and data security.
🎯 Key Topics Covered:
Microsoft 365 Copilot usage and adoption metrics
Viva Insights Copilot Analytics and Dashboard
Microsoft Purview Data Security Posture Management (DSPM) for AI
Measuring AI readiness, impact, and sentiment
Identifying and mitigating risks from third-party Gen AI tools
Shadow IT, oversharing, and compliance risks
Microsoft 365 Admin Center reports and Copilot Readiness
Power BI-based Copilot Business Impact Report (Preview)
📊 Why AI Measurement Matters: Without meaningful measurement, organizations risk operating in the dark—unable to prove ROI, identify friction points, or detect compliance violations. Nikki presents a unified framework combining quantitative metrics, qualitative insights, and risk monitoring to help organizations:
Prove ROI on AI investments
Drive responsible adoption
Protect sensitive data
Ensure compliance and governance
🔍 Tools and Reports Highlighted:
Microsoft 365 Admin Center: Copilot Overview, Usage, Readiness, Agents, Chat, and Adoption Score
Viva Insights Copilot Dashboard: Readiness, Adoption, Impact, Sentiment
Copilot Business Impact Report: Power BI integration for business outcome mapping
Microsoft Purview DSPM for AI: Discover and govern Copilot and third-party Gen AI usage
🔐 Security and Compliance Insights: Learn how to detect unsanctioned Gen AI tools like ChatGPT, Gemini, and Claude, track oversharing, and apply eDLP and Insider Risk Management (IRM) policies. Understand how to use Microsoft Purview—even without E5 Compliance—to monitor Copilot usage and protect sensitive data.
📈 Who Should Watch: This session is ideal for IT leaders, security professionals, compliance officers, and Microsoft 365 admins looking to:
Maximize the value of Microsoft Copilot
Build a secure, measurable AI strategy
Align AI usage with business goals and compliance requirements
🔗 Read the blog https://nikkichapple.com/measuring-copilot-gen-ai/
Master tester AI toolbox - Kari Kakkonen at Testaus ja AI 2025 ProfessioKari Kakkonen
My slides at Professio Testaus ja AI 2025 seminar in Espoo, Finland.
Deck in English, even though I talked in Finnish this time, in addition to chairing the event.
I discuss the different motivations for testing to use AI tools to help in testing, and give several examples in each categories, some open source, some commercial.
Introducing the OSA 3200 SP and OSA 3250 ePRCAdtran
Adtran's latest Oscilloquartz solutions make optical pumping cesium timing more accessible than ever. Discover how the new OSA 3200 SP and OSA 3250 ePRC deliver superior stability, simplified deployment and lower total cost of ownership. Built on a shared platform and engineered for scalable, future-ready networks, these models are ideal for telecom, defense, metrology and more.
European Accessibility Act & Integrated Accessibility TestingJulia Undeutsch
Emma Dawson will guide you through two important topics in this session.
Firstly, she will prepare you for the European Accessibility Act (EAA), which comes into effect on 28 June 2025, and show you how development teams can prepare for it.
In the second part of the webinar, Emma Dawson will explore with you various integrated testing methods and tools that will help you improve accessibility during the development cycle, such as Linters, Storybook, Playwright, just to name a few.
Focus: European Accessibility Act, Integrated Testing tools and methods (e.g. Linters, Storybook, Playwright)
Target audience: Everyone, Developers, Testers
Grannie’s Journey to Using Healthcare AI ExperiencesLauren Parr
AI offers transformative potential to enhance our long-time persona Grannie’s life, from healthcare to social connection. This session explores how UX designers can address unmet needs through AI-driven solutions, ensuring intuitive interfaces that improve safety, well-being, and meaningful interactions without overwhelming users.
Evaluation Challenges in Using Generative AI for Science & Technical ContentPaul Groth
Evaluation Challenges in Using Generative AI for Science & Technical Content.
Foundation Models show impressive results in a wide-range of tasks on scientific and legal content from information extraction to question answering and even literature synthesis. However, standard evaluation approaches (e.g. comparing to ground truth) often don't seem to work. Qualitatively the results look great but quantitive scores do not align with these observations. In this talk, I discuss the challenges we've face in our lab in evaluation. I then outline potential routes forward.
Introduction and Background:
Study Overview and Methodology: The study analyzes the IT market in Israel, covering over 160 markets and 760 companies/products/services. It includes vendor rankings, IT budgets, and trends from 2025-2029. Vendors participate in detailed briefings and surveys.
Vendor Listings: The presentation lists numerous vendors across various pages, detailing their names and services. These vendors are ranked based on their participation and market presence.
Market Insights and Trends: Key insights include IT market forecasts, economic factors affecting IT budgets, and the impact of AI on enterprise IT. The study highlights the importance of AI integration and the concept of creative destruction.
Agentic AI and Future Predictions: Agentic AI is expected to transform human-agent collaboration, with AI systems understanding context and orchestrating complex processes. Future predictions include AI's role in shopping and enterprise IT.
2. 320 Computer Science & Information Technology (CS & IT)
Various methods have been developed for the recognition of handwritten characters, such as the
compressed Column-wise Segmentation of Image Matrix (CSIM) [2] using Neural Network. In
this method, the image matrix is compressed and segmented column-wise then training and
testing using a neural network for different character is performed. The Multi-scale Technique
(MST)[2][3] used for high resolution character sets.
In case of feature based method, Scale Invariant features of characters such as height, width,
centroid, number of bounded regions and textual features such as histogram information are used
for character recognition. The features are feed to a Neural Network [4] for training and testing
purpose. Hand written character recognition using Row-wise Segmentation Technique (RST)
approach used to find out common features along the rows of same characters written in different
hand writing styles and segmenting the matrix into separate rows and finding common rows
among different hand writing styles[5]. Block-wise segmentation technique is also used for the
character recognition [6] by matching the similarity among blocks of the characters.
The complex character such as Hindi, Oriya and Bangla character recognition [7][8] is more
challenging at it produces very alike feature matrices for different characters due to their
structural complexity. Hybrid methods are also applied to recognize the hand written characters.
One such method is a prototype learning/matching method that can be combined with support
vector machines (SVM) in pattern recognition [9].
It is observed that the finding the ideal feature set for particular language and normalizing the
feature matrix is not easy, it requires substantial amount of processor time. In this work we have
done a comparative study on character recognition using feature matrix with a Back Propagation
Neural network (BPN) verses the character recognition using reduced dimensional block
matrix(8x8) with a Perceptron Neural network for English hand written characters i.e. the upper
case, lower case alphabet as well as digits.
2. METHODOLOGY
The sample handwritten characters are collected form ten different persons using black gel pen
i.e. 10 samples of each letter each having different style. These blocks of characters were
digitized using a scanner. Then each character is extracted from the scanned image automatically
and saved with an appropriate name.
3. Computer Science & Information Technology (CS & IT)
Figure 1:
2.1. Image Pre-processing
The individual character image is
character. These involve various tasks such as (1) Binarization to reproduce the image with 0
(black) or 1(white), (2) Thinning to remove the thickness artefact of the pen used for writing
characters, (3) image dilation to restore the continuity of the image pixels. Figure
inverted binarized data sheet of the set of handwritten characters.
2.2. Approach-1: Low dimensional feature based recognition
The images are resized into uniq
dimension. They are converted to a reduced dimensional image matrix of size 8x8. It preserves
only the highly significant features of the character that are used for the character recognition.
The input vector i,e. a [64x1] matrix is prepared from the 8x8 image. It is
and testing of perceptron neural network.
Computer Science & Information Technology (CS & IT)
Figure 1: The Proposed Model for Analysis
e individual character image is pre-processed to produce a skeletal template of the handwritten
character. These involve various tasks such as (1) Binarization to reproduce the image with 0
(black) or 1(white), (2) Thinning to remove the thickness artefact of the pen used for writing
aracters, (3) image dilation to restore the continuity of the image pixels. Figure
inverted binarized data sheet of the set of handwritten characters.
1: Low dimensional feature based recognition
The images are resized into unique standard since the sample character images are different in
hey are converted to a reduced dimensional image matrix of size 8x8. It preserves
only the highly significant features of the character that are used for the character recognition.
The input vector i,e. a [64x1] matrix is prepared from the 8x8 image. It is used for the training
and testing of perceptron neural network.
321
processed to produce a skeletal template of the handwritten
character. These involve various tasks such as (1) Binarization to reproduce the image with 0
(black) or 1(white), (2) Thinning to remove the thickness artefact of the pen used for writing
aracters, (3) image dilation to restore the continuity of the image pixels. Figure-2 show the
images are different in
hey are converted to a reduced dimensional image matrix of size 8x8. It preserves
only the highly significant features of the character that are used for the character recognition.
used for the training
4. 322 Computer Science & Information Technology (CS & IT)
Figure 2: Processed image with bounding box
2.3. Rescaling of image matrix
Case-1: When the original image matrix is a multiple of 8.
i) Input image matrix
ii) Dimension of the original image is divided by 8 i.e. 64 x 32 will be 8 x 4
iii) Original matrix is split into ‘n’ uniform blocks of new dimension
iv) A uniform block is assigned to ‘1’ if the number of 1’s is greater than or equal to the number
of 0’s. Otherwise, a uniform block is assigned to ‘0’.
Case-2: When the original image matrix is not a multiple of 8.
i) Input image matrix
ii) Dimension of the original image is first converted to the nearest multiple of 8 by appending
dummy (zeros) rows and columns i.e. 60 x 50 will be 64 x 56
iii) Dimension of the revised image is divided by 8
iv) Revised matrix is split into ‘n’ uniform blocks of new dimension
v) A uniform block is assigned to ‘1’ if the number of 1’s is greater than or equal to the
number of 0’s. Otherwise, a uniform block is assigned to ‘0’.
In this way images of different dimensions are resized into an 8 x 8 binary matrix. The columns
of 8 x 8 matrixes are stored in a single column matrix one after other. Likewise, 10 samples of
each 26 characters are considered and transformed it into column of size 64 each. As a result of a
training set of 64 x 260 matrix is obtained. Accordingly the target set of 5 x 260 is generated.
5. Computer Science & Information Technology (CS & IT) 323
2.4. Perceptron based ANN training
An Artificial Neural Network (ANN) is an adaptive computational system, it follows perceptron
learning technique. The input layer consists of 64 neurons that represent one character as input,
the hidden layer consists of 32 neuron, where as the output consists of 7 neurons that represents
pattern of 0 and 1 which maps to an individual character. Each neuron is connected to other
neurons by a link associated with weights. The weight contains information about the input,
which is updated during each epoch.
2.5. Approach-2: Scale Invariant feature based recognition
Aspect Ratio: Height and width of character is obtained. The ratio of height and width remain
approximately same for same person for the different characters.
ܴܣ =
ܮ
ܹ
Where, AR=Aspect Ratio
L=Length of Character
W=Width of Character
Occupancy Ratio: This feature is the ratio of number of pixels which belong to the character to
the total pixels in the character image. This feature provides information about character density.
Number of Horizontal Lines: It’s the number of horizontal lines in a character. It’s found out
using a 3x3 horizontal template matrix.
Number of Vertical Lines: It’s the number of vertical lines in a character. It’s found out using a
3x3 vertical template matrix.
Number of Diagonal Lines: It’s the number of diagonal-1 lines in a character. It’s found out
using a 3x3 diagonal-1 and diagonal-2 template matrix.
Number of Bounded Regions: It’s the number of bounded areas found within a character image.
Number of End Points: End points are defined as those pixels, which have only one neighbor in
its eight way neighborhood. Figure-3 the two end points of the image.
Figure 3: End point in an image
Vertical Center of Gravity: Vertical centre of gravity shows the vertical location of the
character image. Vertical centre of gravity of image is calculated as follow
ܿ݃ሺݒሻ =
∑ .ݕ ܰݕ
∑ ܰݕ
Where, Ny: the number of black pixels in each horizontal line Ly with vertical coordinate y,
Horizontal Center of Gravity: Horizontal centre of gravity shows the horizontal location of the
character image. Horizontal centre of gravity of image calculated as follow:
6. 324 Computer Science & Information Technology (CS & IT)
ܿ݃ሺℎሻ =
∑ .ݔ ܰݔ
∑ ܰݔ
Where, Nx: the number of black pixels in each vertical line Lx with horizontal coordinate x
2.6. BPN based ANN Training
Back-propagation neural network is mostly used for the handwritten character recognition since it
support the real values as input to the network. It is a multi-layer feed-forward network which
consists of an input layer, hidden layer and output layer. The normalized values of the scale-
invariant feature matrix are given as the input for the training, the output is pattern of 0’s and 1’s
which maps to an individual character.
2.7. Testing and Recognition
Separate test sets are prepared for testing purpose. After the training process is completed, the
test pattern is fed to the neural network to check the learning ability of the trained net. The output
of the simulation is compared with the specified target set. The character is recognized by
selective thresholding technique.
3. RESULT ANALYSIS
The OCR system is developed using MATLAB. The experimental results are shown below which
is carried out to recognize the “A” and “5” as the inputs. The Figure-4 represents the normalized
values of the scale invariant features of the English Alpha-numerics. Figure-5 indicates the
performance analysis of BPN network as trained with normalized feature values. The Figure-6
shows the interface to load the test image and to select the training type such as BPN based
training. Figure-7 indicates the recognition of “A”, The Figure-8 shows the LDF matrix, Figure-9
represents the perceptron based training and testing. The recognition of “5” is shown in Figure-
10.
Figure 4: Scale Invariant normalized feature matrix for ‘A’-‘z’ & ‘0’-‘9’
7. Computer Science & Information Technology (CS & IT) 325
Figure 5: BPN based training scale invariant features
Figure 6: User Interface for Load Image
8. 326 Computer Science & Information Technology (CS & IT)
Figure 7: Recognition of character ‘A’ using BPN
Figure 8: Reduced dimensional feature matrix[64x1] for ‘A’-‘z’ & ‘0’-‘9’
9. Computer Science & Information Technology (CS & IT) 327
The target matrix of
‘A’ is
[ 0
0
0
0
1 ]
The output matrix of ‘A’ is
Figure 9: Perception based training & testing
Figure 10: Recognition of character ‘5’ using perceptron network
4. CONCLUSION
In this work it has been observed that finding the reduced dimensional feature matrix of an image
is easy in comparison with the scale invariant feature matrix. The training performance of the SIF
matrix is much faster and reliable. The performance of the network depends upon selection of the
features into the SIF matrix. The feature selection is more challenging in case of structurally
complex scripts such as Bangla, Hindi and Oriya. It has been observed that using the SIF features,
the English alphabets and numeric’s matches with an accuracy of 95% on average matching while
the LDF match accuracy varies from 78% to 96% for different characters.
The work may be further extended to test the regional language scripts. It can be tested with
different reduced dimensional matrix of different dimensions.
10. 328 Computer Science & Information Technology (CS & IT)
REFERENCES
[1] Vijay Patil and Sanjay Shimpi, “Handwritten English Character Recognition Using Neural Network”,
Elixir Comp. Sci. & Engg. 41 (2011) 5587-5591, November 2011.
[2] Velappa Ganapathy, and Kok Leong Liew, “Handwritten Character Recognition Using Multi-scale
Neural Network Training Technique”, World Academy of Science, Engineering and Technology 39
2008.
[3] Rakesh Kumar Mandal, N R Manna, “Hand Written English Character Recognition Using Column-
Wise Segmentation of Image Matrix (CSIM)”, Issue 5, Volume 11, May 2012.
[4] Rakesh Kumar Mandal, N R Manna, “Hand Written English Character Recognition Using Row-Wise
Segmentation (RST)”, International Symposium on Devices MEMS, Intelligent Systems &
Communication (ISDMISC) 2011.
[5] Apash Roy, N. R. Manna, “Handwritten Character Recognition Using Block wise Segmentation
Technique (BST) in Neural Network”.
[6] Soumya Mishra, Debashish Nanda, SanghamitraMohanty, “Oriya character recognition using Neural
Networks”, Special Issue of IJCCT Vol. 2 Issue 2, 3, 4; 2010 for International Conference [ICCT-
2010], 3rd-5th December 2010.
[7] B. Kumar, N. Kumar, C. Palai,.P. K. Jena, S. Chattopadhyay, “Optical Character Recognition using
Ant Miner Algorithm: A Case Study on Oriya Character Recognition”, International Journal of
Computer Applications (0975 – 8887) Volume 61– No.3, January 2013.
[8] Fu Chang, Chin-Chin Linand Chun-Jen Chen,Institute of Information Science, Academia Sinica ,
Taipei, Taiwan Dept. of Electrical Engineering, National Taipei University of Technology, Taipei,
Taiwan”Applying A Hybrid Method To Handwritten Character Recognition”.
AUTHORS
Pradeep Kumar Jena Working as an Associate Professor in the Dept. of MCA at National
Institute of Science and Technology, Berhampur. His domain of research includes Biometric
AuthenticationImage Processing, Pattern Recognition and Data Mining.
Charulata Palai Working as an Assistant Professor in the Dept. of CSE at National Institute
of Science and Technology, Berhampur. Her domain of research includes Soft
ComputingImage Processing, Pattern Recognition.
Lopamudra Sahoo is a BTech Final Year student of NIST, Berhampur.
Anshupa Patel is a BTech Final Year student of NIST, Berhampur.