Nowadays image compression has become a necessity due to a large volume of images. For efficient use of storage space and data transmission, it becomes essential to compress the image. In this paper, we propose a dictionary based image compression framework via sparse representation, with the construction of a trained over-complete dictionary. The overcomplete dictionary is trained using the intra-prediction residuals obtained from different images and is applied for sparse representation. In this method, the current image block is first predicted from its spatially neighboring blocks, and then the prediction residuals are encoded via sparse representation. Sparse approximation algorithm and the trained overcomplete dictionary are applied for sparse representation of prediction residuals. The detail coefficients obtained from sparse representation are used for encoding. Experimental result shows that the proposed method yields both improved coding efficiency and image quality as compared to some state-of-the-art image compression methods.
Documents Image Binarization is performed in the preprocessing stage for document analysis and it aims to
segment the foreground text from the document background. A fast and accurate document image binarization
technique is important for the ensuing document image processing tasks such as optical character recognition
(OCR). Though document image binarization has been studied for many years, the thresholding of degraded
document images is still an unsolved problem due to the high inter/intra variation between the text stroke and
the document background across different document images. The handwritten text within the degraded
documents often shows a certain amount of variation in terms of the stroke width, stroke brightness, stroke
connection, and document background. In addition, historical documents are often degraded by the bleed.
Documents are often degraded by different types of imaging artifact. These different types of document
degradations tend to induce the document thresholding error and make degraded document image binarization a
big challenge to most state-of-the-art techniques. The proposed method is simple, robust and capable of
handling different types of degraded document images with minimum parameter tuning. It makes use of the
adaptive image contrast that combines the local image contrast and the local image gradient adaptively and
therefore is tolerant to the text and background variation caused by different types of document degradations. In
particular, the proposed technique addresses the over-normalization problem of the local maximum minimum
algorithm. At the same time, the parameters used in the algorithm can be adaptively estimated.
This document presents a scalable method for image classification using sparse coding and dictionary learning. It proposes parallelizing the computation of image similarity for faster recognition. Specifically, it distributes the task of measuring similarity between images among multiple cores in a cluster. Experimental results on a face recognition dataset show nearly linear speedup when balancing the dataset size and number of nodes. Reconstruction errors are used as a similarity measure, with dictionaries learned using K-SVD for each image. The proposed parallel method distributes this similarity computation process to achieve faster image classification.
Fuzzy Encoding For Image Classification Using Gustafson-Kessel AglorithmAshish Gupta
This paper presents a novel adaptation of fuzzy clustering and
feature encoding for image classification. Visual word ambiguity
has recently been successfully modeled by kernel codebooks
to provide improvement in classification performance
over the standard ‘Bag-of-Features’(BoF) approach, which
uses hard partitioning and crisp logic for assignment of features
to visual words. Motivated by this progress we utilize
fuzzy logic to model the ambiguity and combine it with clustering
to discover fuzzy visual words. The feature descriptors
of an image are encoded using the learned fuzzy membership
function associated with each word. The codebook built
using this fuzzy encoding technique is demonstrated to provide
superior performance over BoF. We use the Gustafson-
Kessel algorithm which is an improvement over Fuzzy CMeans
clustering and can adapt to local distributions. We
evaluate our approach on several popular datasets and demonstrate
that it consistently provides superior performance to the
BoF approach.
Comparative Analysis of Lossless Image Compression Based On Row By Row Classi...IJERA Editor
This document proposes and evaluates a near lossless image compression algorithm that divides color images into red, green, and blue channels. It classifies pixels in each channel row-by-row and records the results in mask images. The image data is then decomposed into sequences based on the classification and the mask images are hidden in the least significant bits of the sequences. Different encoding schemes like LZW, Huffman, and RLE are applied and compared. Experimental results on test images show the proposed algorithm achieves smaller bits per pixel than simple encoding schemes. PSNR values also indicate very little difference between original and reconstructed images.
Improved block based segmentation for jpeg compressed document imageseSAT Journals
Abstract
Image Compression is to minimize the size in bytes of a graphics file without degrading the quality of the image to an unacceptable
level. The compound image compression normally based on three classification methods that is object based, layer based and block
based. This paper presents a block-based segmentation. for visually lossless compression of scanned documents that contain not only
photographic images but also text and graphic images. In low bit rate applications they suffer with undesirable compression artifacts,
especially for document images. Existing methods can reduce these artifacts by using post processing methods without changing the
encoding process. Some of these post processing methods requires classification of the encoded blocks into different categories.
Keywords- AC energy, Discrete Cosine Transform (DCT), JPEG, K-means clustering, Threshold value
This document discusses techniques for representing digital circuit partitioning problems using graph representations. It presents three encoding techniques to map graph partitions to the problem domain: 1) a binary string where each bit represents a cell and its partition, 2) a string with two regions to represent vertices and edge crossings, and 3) a string with regions for vertices and edges. The techniques are evaluated in terms of suitability, with the second approach more suitable for dense circuits. Net cut evaluation is also described to analyze partitioning solutions.
Halftoning-based BTC image reconstruction using patch processing with border ...TELKOMNIKA JOURNAL
This paper presents a new halftoning-based block truncation coding (HBTC) image reconstruction using sparse representation framework. The HBTC is a simple yet powerful image compression technique, which can effectively remove the typical blocking effect and false contour. Two types of HBTC methods are discussed in this paper, i.e., ordered dither block truncation coding (ODBTC) and error diffusion block truncation coding (EDBTC). The proposed sparsity-based method suppresses the impulsive noise on ODBTC and EDBTC decoded image with a coupled dictionary containing the HBTC image component and the clean image component dictionaries. Herein, a sparse coefficient is estimated from the HBTC decoded image by means of the HBTC image dictionary. The reconstructed image is subsequently built and aligned from the clean, i.e. non-compressed image dictionary and predicted sparse coefficient. To further reduce the blocking effect, the image patch is firstly identified as “border” and “non-border” type before applying the sparse representation framework. Adding the Laplacian prior knowledge on HBTC decoded image, it yields better reconstructed image quality. The experimental results demonstrate the effectiveness of the proposed HBTC image reconstruction. The proposed method also outperforms the former schemes in terms of reconstructed image quality.
Perimetric Complexity of Binary Digital ImagesRSARANYADEVI
Perimetric complexity is a measure of the complexity of binary pictures. It is defined as the sum of inside and outside perimeters of the foreground, squared, divided by the foreground area, divided by . Difficulties arise when this definition is applied to digital images composed of binary pixels. In this article we identify these problems and propose solutions. Perimetric complexity is often used as a measure of visual complexity, in which case it should take into account the limited resolution of the visual system. We propose a measure of visual perimetric complexity that meets this requirement.
The document reviews approaches to image interpolation and super-resolution. It discusses several interpolation methods including polynomial-based, edge-directed, and soft-decision approaches. Edge-directed methods aim to preserve edge sharpness during upsampling by estimating edge orientations or fusing multiple orientations. New edge-directed interpolation uses a Wiener filter to estimate missing pixel values. Soft-decision adaptive interpolation and robust soft-decision interpolation further improve results by modeling image signals within local windows and incorporating outlier weighting. The document provides formulations and comparisons of these methods.
This document proposes an efficient data steganography method called Adaptive Pixel Pair Matching (APPM) with high security. APPM hides data by substituting pixel pairs in a cover image based on a secret key. It defines an extraction function and compact neighborhood set for pixel pairs to minimize embedding distortion. APPM converts the secret message into digits of a B-ary numerical system for hiding. It calculates the optimal value of B and neighborhood set based on the image and message size. APPM generates a random embedding sequence using a key for substitution. It also provides an external password for additional security of the hidden message. The document claims this method provides better image quality and higher payload than previous pixel pair matching methods with increased security.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
An improved image compression algorithm based on daubechies wavelets with ar...Alexander Decker
This document summarizes an academic article that proposes a new image compression algorithm using Daubechies wavelets and arithmetic coding. It first discusses existing image compression techniques and their limitations. It then describes the proposed algorithm, which applies Daubechies wavelet transform followed by 2D Walsh wavelet transform on image blocks and arithmetic coding. Results show the proposed method achieves higher compression ratios and PSNR values than existing algorithms like EZW and SPIHT. Future work aims to improve results by exploring different wavelets and compression techniques.
This is a paper I wrote as part of my seminar "Inverse Problems in Computer Vision" while pursuing my M.Sc Medical Engineering at FAU, Erlangen, Germany.
The paper details a state-of-the-art method used for Single Image Super Resolution using Deep Convolutional Networks and the possible extensions to the original approach by considering compression and noise artifacts.
A systematic image compression in the combination of linear vector quantisati...eSAT Publishing House
1) The document presents a method for image compression that combines linear vector quantization and discrete wavelet transform.
2) Linear vector quantization is used to generate codebooks and encode image blocks, achieving better PSNR and MSE than self-organizing maps.
3) The encoded blocks are then subjected to discrete wavelet transform. Low-low subbands are stored for reconstruction while other subbands are discarded.
4) Experimental results show the proposed method achieves higher PSNR and lower MSE than existing techniques, preserving both texture and edge information.
The document discusses appearance-based face recognition using PCA and LDA algorithms. It summarizes the steps of each algorithm and compares their performance on preprocessed face images from the Faces94 database. Image preprocessing techniques like grayscale conversion and modified histogram equalization are applied before PCA and LDA to enhance image quality and improve recognition rates. The paper aims to study PCA and LDA with respect to recognition accuracy and dimensionality.
This document discusses fractal image compression based on jointly and different partitioning schemes. It proposes partitioning RGB images into range blocks in two ways: 1) Jointly, where the red, green, and blue channels are partitioned together into blocks of the same size and coordinates. 2) Differently, where each channel is partitioned independently, resulting in different block sizes and coordinates for each channel. The document provides background on fractal image compression and the encoding/decoding processes. It analyzes the two partitioning schemes and argues the different scheme is more effective for encoding by allowing each channel to have customized partitioning.
The document discusses a method for compressing color images using block truncation coding (BTC) and genetic algorithms. BTC works by dividing images into blocks and quantizing each block to a high or low value based on the block's mean. This reduces quality issues with BTC. Color images have correlated red, green, and blue planes. The method uses a common bit plane optimized with genetic algorithms to represent all three color planes, improving quality and compression ratio compared to standard BTC and error diffused BTC. Experimental results showed the proposed method provided higher quality reconstructed images as measured by peak signal-to-noise ratio.
Zernike moment of invariants for effective image retrieval using gaussian fil...IAEME Publication
This document summarizes a research paper that proposes using Zernike moments and steerable Gaussian filters for effective image retrieval based on color, texture, and shape features. It describes extracting dominant colors from images using dynamic color quantization and clustering. Texture features are represented using steerable filter decomposition. Shape features are described using pseudo-Zernike moments. The proposed method combines these color, texture, and shape features to generate a robust feature set for image retrieval that provides better retrieval accuracy than other methods.
Hierarchical Approach for Total Variation Digital Image InpaintingIJCSEA Journal
The art of recovering an image from damage in an undetectable form is known as inpainting. The manual work of inpainting is most often a very time consum ing process. Due to digitalization of this technique, it is automatic and faster. In this paper, after the user selects the regions to be reconstructed, the algorithm automatically reconstruct the lost regions with the help of the information surrounding them. The existing methods perform very well when the region to be reconstructed is very small, but fails in proper reconstruction as the area increases. This paper describes a Hierarchical method by which the area to be inpainted is reduced in multiple levels and Total Variation(TV) method is used to inpaint in each level. This algorithm gives better performance when compared with other existing algorithms such as nearest neighbor interpolation, Inpainting through Blurring and Sobolev Inpainting.
The document describes an algorithm for detecting text in camera-captured images. It begins with preprocessing steps like converting the color image to grayscale, applying edge detection and morphological operations like dilation and erosion. This gives initial bounding boxes containing candidate text regions. Further processing includes applying geometrical constraints to filter boxes, performing multiresolution analysis, connected component analysis and filtering by area to get the final text regions. Inversion and addition steps are used to handle text against different backgrounds.
Modified Skip Line Encoding for Binary Image Compressionidescitation
This paper proposes a modified skip line encoding technique for lossless compression of binary images. Skip line encoding exploits correlation between successive scan lines by encoding only one line and skipping similar lines. The proposed technique improves upon existing skip line encoding by allowing a scan line to be skipped if a similar line exists anywhere in the image, rather than just successive lines. Experimental results on sample images show the modified technique achieves higher compression ratios than conventional skip line encoding.
GREY LEVEL CO-OCCURRENCE MATRICES: GENERALISATION AND SOME NEW FEATURESijcseit
Grey Level Co-occurrence Matrices (GLCM) are one of the earliest techniques used for image texture
analysis. In this paper we defined a new feature called trace extracted from the GLCM and its implications
in texture analysis are discussed in the context of Content Based Image Retrieval (CBIR). The theoretical
extension of GLCM to n-dimensional gray scale images are also discussed. The results indicate that trace
features outperform Haralick features when applied to CBIR.
Image Super-Resolution using Single Image Semi Coupled Dictionary Learning CSCJournals
Obtaining a high resolution image from a low resolution image plays an important role in many
image processing applications. In Single Image Super Resolution (SISR), the desired high
resolution output image is synthesized from a single low resolution input image. In this paper,
Single Image Semi Coupled Dictionary Learning (SI-SCDL) method is proposed, where the
dictionaries to represent the high and low resolution images are trained from the input image
itself. In the proposed method, the online training stage is employed, where the dictionaries are
learnt online and it does not require any external training database. Simulation results show that
the proposed SI-SCDL method performs better when compared to other mentioned methods.
This document discusses using a Direction-Length Code (DLC) to represent binary objects. The DLC is a "knowledge vector" that provides information about the direction and length of pixels in every direction of an object. Patterns over a 3x3 pixel array are generated to form a basic alphabet for representing digital images as spatial distributions of these patterns. The DLC compresses bi-level images while preserving shape information and allowing significant data reduction. It can serve as standard input for numerous shape analysis algorithms. Components of images are extracted from the DLC and used to accurately regenerate the original images, demonstrating the effectiveness of the DLC representation.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
Image Inpainting Using Cloning AlgorithmsCSCJournals
In image recovery image inpainting has become essential content and crucial topic in research of a new era. The objective is to restore the image with the surrounding information or modifying an image in a way that looks natural for the viewer. The process involves transporting and diffusing image information. In this paper to inpaint an image cloning concept has been used. Multiscale transformation method is used for cloning process of an image inpainting. Results are compared with conventional methods namely Taylor expansion method, poisson editing, Shepard’s method. Experimental analysis verifies better results and shows that Shepard’s method using multiscale transformation not only restores small scale damages but also large damaged area and useful in duplication of image information in an image.
SECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSIONsipij
In this paper, we propose a secure Orthogonal Matching Pursuit (OMP) based pattern recognition scheme that well supports image compression. The secure OMP is a sparse coding algorithm that chooses atoms sequentially and calculates sparse coefficients from encrypted images. The encryption is carried out by using a random unitary transform. The proposed scheme offers two prominent features. 1) It is capable of
pattern recognition that works in the encrypted image domain. Even if data leaks, privacy can be maintained because data remains encrypted. 2) It realizes Encryption-then-Compression (EtC) systems, where image encryption is conducted prior to compression. The pattern recognition can be carried out using a
few sparse coefficients. On the basis of the pattern recognition results, the scheme can compress selected images with high quality by estimating a sufficient number of sparse coefficients. We use the INRIA dataset to demonstrate its performance in detecting humans in images. The proposal is shown to realize human detection with encrypted images and efficiently compress the images selected in the image recognition stage.
K-SVD: ALGORITHM FOR FINGERPRINT COMPRESSION BASED ON SPARSE REPRESENTATION ijiert bestjournal
In current years there has been an increasing interest in the study of sparse representation of
signals. Using an overcomplete glossary that contains prototype signal-atoms, signals are
described by sparse linear combinations of these atoms. Recognition of persons by means of
biometric description is an important technology in the society, because biometric identifiers
cannot be shared and they intrinsically characterize the individual’s bodily distinctiveness.
Among several biometric recognition technologies, fingerprint compression is very popular
for personal identification. One more fingerprint compression algorithm based on sparse
representation using K-SVD algorithm is introduced. In the algorithm, First we construct a
dictionary for predefined fingerprint photocopy patches. For a new given fingerprint images,
suggest its patches according to the dictionary by computing
-minimization by MP method
and then quantize and encode the representation.This paper comparesdissimilarcompression
standards like JPEG,JPEG-2000,WSQ,K-SVDetc. The paper show that this is effective
compared with several competing compression techniques particularly at high compression
ratios.
The document summarizes an efficient image compression technique using Overlapped Discrete Cosine Transform (MDCT) combined with adaptive thinning.
In the first phase, MDCT is applied which is based on DCT-IV but with overlapping blocks, enabling robust compression. In the second phase, adaptive thinning recursively removes points from the image based on Delaunay triangulations, further compressing the image. Simulation results showed over 80% pixel reduction with 30dB PSNR, requiring less points for the compressed image. The technique combines MDCT for frequency-domain compression with adaptive thinning for spatial-domain compression.
The document reviews approaches to image interpolation and super-resolution. It discusses several interpolation methods including polynomial-based, edge-directed, and soft-decision approaches. Edge-directed methods aim to preserve edge sharpness during upsampling by estimating edge orientations or fusing multiple orientations. New edge-directed interpolation uses a Wiener filter to estimate missing pixel values. Soft-decision adaptive interpolation and robust soft-decision interpolation further improve results by modeling image signals within local windows and incorporating outlier weighting. The document provides formulations and comparisons of these methods.
This document proposes an efficient data steganography method called Adaptive Pixel Pair Matching (APPM) with high security. APPM hides data by substituting pixel pairs in a cover image based on a secret key. It defines an extraction function and compact neighborhood set for pixel pairs to minimize embedding distortion. APPM converts the secret message into digits of a B-ary numerical system for hiding. It calculates the optimal value of B and neighborhood set based on the image and message size. APPM generates a random embedding sequence using a key for substitution. It also provides an external password for additional security of the hidden message. The document claims this method provides better image quality and higher payload than previous pixel pair matching methods with increased security.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
An improved image compression algorithm based on daubechies wavelets with ar...Alexander Decker
This document summarizes an academic article that proposes a new image compression algorithm using Daubechies wavelets and arithmetic coding. It first discusses existing image compression techniques and their limitations. It then describes the proposed algorithm, which applies Daubechies wavelet transform followed by 2D Walsh wavelet transform on image blocks and arithmetic coding. Results show the proposed method achieves higher compression ratios and PSNR values than existing algorithms like EZW and SPIHT. Future work aims to improve results by exploring different wavelets and compression techniques.
This is a paper I wrote as part of my seminar "Inverse Problems in Computer Vision" while pursuing my M.Sc Medical Engineering at FAU, Erlangen, Germany.
The paper details a state-of-the-art method used for Single Image Super Resolution using Deep Convolutional Networks and the possible extensions to the original approach by considering compression and noise artifacts.
A systematic image compression in the combination of linear vector quantisati...eSAT Publishing House
1) The document presents a method for image compression that combines linear vector quantization and discrete wavelet transform.
2) Linear vector quantization is used to generate codebooks and encode image blocks, achieving better PSNR and MSE than self-organizing maps.
3) The encoded blocks are then subjected to discrete wavelet transform. Low-low subbands are stored for reconstruction while other subbands are discarded.
4) Experimental results show the proposed method achieves higher PSNR and lower MSE than existing techniques, preserving both texture and edge information.
The document discusses appearance-based face recognition using PCA and LDA algorithms. It summarizes the steps of each algorithm and compares their performance on preprocessed face images from the Faces94 database. Image preprocessing techniques like grayscale conversion and modified histogram equalization are applied before PCA and LDA to enhance image quality and improve recognition rates. The paper aims to study PCA and LDA with respect to recognition accuracy and dimensionality.
This document discusses fractal image compression based on jointly and different partitioning schemes. It proposes partitioning RGB images into range blocks in two ways: 1) Jointly, where the red, green, and blue channels are partitioned together into blocks of the same size and coordinates. 2) Differently, where each channel is partitioned independently, resulting in different block sizes and coordinates for each channel. The document provides background on fractal image compression and the encoding/decoding processes. It analyzes the two partitioning schemes and argues the different scheme is more effective for encoding by allowing each channel to have customized partitioning.
The document discusses a method for compressing color images using block truncation coding (BTC) and genetic algorithms. BTC works by dividing images into blocks and quantizing each block to a high or low value based on the block's mean. This reduces quality issues with BTC. Color images have correlated red, green, and blue planes. The method uses a common bit plane optimized with genetic algorithms to represent all three color planes, improving quality and compression ratio compared to standard BTC and error diffused BTC. Experimental results showed the proposed method provided higher quality reconstructed images as measured by peak signal-to-noise ratio.
Zernike moment of invariants for effective image retrieval using gaussian fil...IAEME Publication
This document summarizes a research paper that proposes using Zernike moments and steerable Gaussian filters for effective image retrieval based on color, texture, and shape features. It describes extracting dominant colors from images using dynamic color quantization and clustering. Texture features are represented using steerable filter decomposition. Shape features are described using pseudo-Zernike moments. The proposed method combines these color, texture, and shape features to generate a robust feature set for image retrieval that provides better retrieval accuracy than other methods.
Hierarchical Approach for Total Variation Digital Image InpaintingIJCSEA Journal
The art of recovering an image from damage in an undetectable form is known as inpainting. The manual work of inpainting is most often a very time consum ing process. Due to digitalization of this technique, it is automatic and faster. In this paper, after the user selects the regions to be reconstructed, the algorithm automatically reconstruct the lost regions with the help of the information surrounding them. The existing methods perform very well when the region to be reconstructed is very small, but fails in proper reconstruction as the area increases. This paper describes a Hierarchical method by which the area to be inpainted is reduced in multiple levels and Total Variation(TV) method is used to inpaint in each level. This algorithm gives better performance when compared with other existing algorithms such as nearest neighbor interpolation, Inpainting through Blurring and Sobolev Inpainting.
The document describes an algorithm for detecting text in camera-captured images. It begins with preprocessing steps like converting the color image to grayscale, applying edge detection and morphological operations like dilation and erosion. This gives initial bounding boxes containing candidate text regions. Further processing includes applying geometrical constraints to filter boxes, performing multiresolution analysis, connected component analysis and filtering by area to get the final text regions. Inversion and addition steps are used to handle text against different backgrounds.
Modified Skip Line Encoding for Binary Image Compressionidescitation
This paper proposes a modified skip line encoding technique for lossless compression of binary images. Skip line encoding exploits correlation between successive scan lines by encoding only one line and skipping similar lines. The proposed technique improves upon existing skip line encoding by allowing a scan line to be skipped if a similar line exists anywhere in the image, rather than just successive lines. Experimental results on sample images show the modified technique achieves higher compression ratios than conventional skip line encoding.
GREY LEVEL CO-OCCURRENCE MATRICES: GENERALISATION AND SOME NEW FEATURESijcseit
Grey Level Co-occurrence Matrices (GLCM) are one of the earliest techniques used for image texture
analysis. In this paper we defined a new feature called trace extracted from the GLCM and its implications
in texture analysis are discussed in the context of Content Based Image Retrieval (CBIR). The theoretical
extension of GLCM to n-dimensional gray scale images are also discussed. The results indicate that trace
features outperform Haralick features when applied to CBIR.
Image Super-Resolution using Single Image Semi Coupled Dictionary Learning CSCJournals
Obtaining a high resolution image from a low resolution image plays an important role in many
image processing applications. In Single Image Super Resolution (SISR), the desired high
resolution output image is synthesized from a single low resolution input image. In this paper,
Single Image Semi Coupled Dictionary Learning (SI-SCDL) method is proposed, where the
dictionaries to represent the high and low resolution images are trained from the input image
itself. In the proposed method, the online training stage is employed, where the dictionaries are
learnt online and it does not require any external training database. Simulation results show that
the proposed SI-SCDL method performs better when compared to other mentioned methods.
This document discusses using a Direction-Length Code (DLC) to represent binary objects. The DLC is a "knowledge vector" that provides information about the direction and length of pixels in every direction of an object. Patterns over a 3x3 pixel array are generated to form a basic alphabet for representing digital images as spatial distributions of these patterns. The DLC compresses bi-level images while preserving shape information and allowing significant data reduction. It can serve as standard input for numerous shape analysis algorithms. Components of images are extracted from the DLC and used to accurately regenerate the original images, demonstrating the effectiveness of the DLC representation.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
Image Inpainting Using Cloning AlgorithmsCSCJournals
In image recovery image inpainting has become essential content and crucial topic in research of a new era. The objective is to restore the image with the surrounding information or modifying an image in a way that looks natural for the viewer. The process involves transporting and diffusing image information. In this paper to inpaint an image cloning concept has been used. Multiscale transformation method is used for cloning process of an image inpainting. Results are compared with conventional methods namely Taylor expansion method, poisson editing, Shepard’s method. Experimental analysis verifies better results and shows that Shepard’s method using multiscale transformation not only restores small scale damages but also large damaged area and useful in duplication of image information in an image.
SECURE OMP BASED PATTERN RECOGNITION THAT SUPPORTS IMAGE COMPRESSIONsipij
In this paper, we propose a secure Orthogonal Matching Pursuit (OMP) based pattern recognition scheme that well supports image compression. The secure OMP is a sparse coding algorithm that chooses atoms sequentially and calculates sparse coefficients from encrypted images. The encryption is carried out by using a random unitary transform. The proposed scheme offers two prominent features. 1) It is capable of
pattern recognition that works in the encrypted image domain. Even if data leaks, privacy can be maintained because data remains encrypted. 2) It realizes Encryption-then-Compression (EtC) systems, where image encryption is conducted prior to compression. The pattern recognition can be carried out using a
few sparse coefficients. On the basis of the pattern recognition results, the scheme can compress selected images with high quality by estimating a sufficient number of sparse coefficients. We use the INRIA dataset to demonstrate its performance in detecting humans in images. The proposal is shown to realize human detection with encrypted images and efficiently compress the images selected in the image recognition stage.
K-SVD: ALGORITHM FOR FINGERPRINT COMPRESSION BASED ON SPARSE REPRESENTATION ijiert bestjournal
In current years there has been an increasing interest in the study of sparse representation of
signals. Using an overcomplete glossary that contains prototype signal-atoms, signals are
described by sparse linear combinations of these atoms. Recognition of persons by means of
biometric description is an important technology in the society, because biometric identifiers
cannot be shared and they intrinsically characterize the individual’s bodily distinctiveness.
Among several biometric recognition technologies, fingerprint compression is very popular
for personal identification. One more fingerprint compression algorithm based on sparse
representation using K-SVD algorithm is introduced. In the algorithm, First we construct a
dictionary for predefined fingerprint photocopy patches. For a new given fingerprint images,
suggest its patches according to the dictionary by computing
-minimization by MP method
and then quantize and encode the representation.This paper comparesdissimilarcompression
standards like JPEG,JPEG-2000,WSQ,K-SVDetc. The paper show that this is effective
compared with several competing compression techniques particularly at high compression
ratios.
The document summarizes an efficient image compression technique using Overlapped Discrete Cosine Transform (MDCT) combined with adaptive thinning.
In the first phase, MDCT is applied which is based on DCT-IV but with overlapping blocks, enabling robust compression. In the second phase, adaptive thinning recursively removes points from the image based on Delaunay triangulations, further compressing the image. Simulation results showed over 80% pixel reduction with 30dB PSNR, requiring less points for the compressed image. The technique combines MDCT for frequency-domain compression with adaptive thinning for spatial-domain compression.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
ECCV2010: feature learning for image classification, part 2zukun
The document discusses sparse coding, an unsupervised machine learning technique for image representation. Sparse coding learns a dictionary of basic image features called bases from unlabeled image data. It then represents each image as a sparse linear combination of the bases. This produces a more compact representation than raw pixels and interprets images as combinations of basic visual concepts like edges. The technique was inspired by representations in the visual cortex and can be combined with features like SIFT for improved performance.
A Compressed Sensing Approach to Image Reconstructionijsrd.com
compressed sensing is a new technique that discards the Shannon Nyquist theorem for reconstructing a signal. It uses very few random measurements that were needed traditionally to recover any signal or image. The need of this technique comes from the fact that most of the information is provided by few of the signal coefficients, then why do we have to acquire all the data if it is thrown away without being used. A number of review articles and research papers have been published in this area. But with the increasing interest of practitioners in this emerging field it is mandatory to take a fresh look at this method and its implementations. The main aim of this paper is to review the compressive sensing theory and its applications.
A spatial image compression algorithm based on run length encodingjournalBEEI
Image compression is vital for many areas such as communication and storage of data that is rapidly growing nowadays. In this paper, a spatial lossy compression algorithm for gray scale images is presented. It exploits the inter-pixel and the psycho-visual data redundancies in images. The proposed technique finds paths of connected pixels that fluctuate in value within some small threshold. The path is calculated by looking at the 4-neighbors of a pixel then choosing the best one based on two conditions; the first is that the selected pixel must not be included in another path and the second is that the difference between the first pixel in the path and the selected pixel is within the specified threshold value. A path starts with a given pixel and consists of the locations of the subsequently selected pixels. Run-length encoding scheme is applied on paths to harvest the inter-pixel redundancy. After applying the proposed algorithm on several test images, a promising quality vs. compression ratio results have been achieved.
Many algorithms have been developed to find sparse representation over redundant dictionaries or
transform. This paper presents a novel method on compressive sensing (CS)-based image compression
using sparse basis on CDF9/7 wavelet transform. The measurement matrix is applied to the three levels of
wavelet transform coefficients of the input image for compressive sampling. We have used three different
measurement matrix as Gaussian matrix, Bernoulli measurement matrix and random orthogonal matrix.
The orthogonal matching pursuit (OMP) and Basis Pursuit (BP) are applied to reconstruct each level of
wavelet transform separately. Experimental results demonstrate that the proposed method given better
quality of compressed image than existing methods in terms of proposed image quality evaluation indexes
and other objective (PSNR/UIQI/SSIM) measurements.
Fingerprint Image Compression using Sparse Representation and Enhancement wit...Editor IJCATR
A technique for enhancing decompressed fingerprint image using Wiener2 filter is proposed. First compression is done by sparse representation. Compression of fingerprint is necessary for reducing the memory consumption and efficient transfer of fingerprint images. This is very essential for the application which includes access control and forensics. So the fingerprint image is compressed using sparse representation. In this technique, first dictionary is constructed for patches of fingerprint images. Then a fingerprint is selected and the coefficients are obtained and encoded. Thus the compressed fingerprint is obtained. But when the fingerprint is reconstructed, it is affected by noise. So Wiener2 filter is used to filter the noise in the image. The ridge and bifurcation count is extracted from decompressed and enhanced fingerprints. The experiment result shows that the enhanced fingerprint image preserves more bifurcation than decompressed fingerprint image. The future analysis can be considered for preserving ridges.
Comparison of different Fingerprint Compression Techniquessipij
The important features of wavelet transform and different methods in compression of fingerprint images have been implemented. Image quality is measured objectively using peak signal to noise ratio (PSNR) and mean square error (MSE).A comparative study using discrete cosine transform based Joint Photographic Experts Group(JPEG) standard , wavelet based basic Set Partitioning in Hierarchical trees(SPIHT) and Modified SPIHT is done. The comparison shows that Modified SPIHT offers better compression than basic SPIHT and JPEG. The results will help application developers to choose a good wavelet compression system for their applications.
Performance analysis of compressive sensing recovery algorithms for image pr...IJECEIAES
The modern digital world comprises of transmitting media files like image, audio, and video which leads to usage of large memory storage, high data transmission rate, and a lot of sensory devices. Compressive sensing (CS) is a sampling theory that compresses the signal at the time of acquiring it. Compressive sensing samples the signal efficiently below the Nyquist rate to minimize storage and recoveries back the signal significantly minimizing the data rate and few sensors. The proposed paper proceeds with three phases. The first phase describes various measurement matrices like Gaussian matrix, circulant matrix, and special random matrices which are the basic foundation of compressive sensing technique that finds its application in various fields like wireless sensors networks (WSN), internet of things (IoT), video processing, biomedical applications, and many. Finally, the paper analyses the performance of the various reconstruction algorithms of compressive sensing like basis pursuit (BP), compressive sampling matching pursuit (CoSaMP), iteratively reweighted least square (IRLS), iterative hard thresholding (IHT), block processing-based basis pursuit (BP-BP) based on mean square error (MSE), and peak signal to noise ratio (PSNR) and then concludes with future works.
Compression can be defined as an art form that involves the representation of information in a reduced form when compared to the original information. Image compression is extremely important in this day and age because of the increased demand for sharing and storing multimedia data. Compression is concerned with removing redundant or superfluous information from a file to reduce the size of the file. The reduction of the file size saves both memory and the time required to transmit and store data. Lossless compression techniques are distinguished from lossy compression techniques, which are distinguished from one another. This paper focuses on the literature studies on various compression techniques and the comparisons between them.
Performance Analysis on Fingerprint Image Compression Using K-SVD-SR and SPIHTIRJET Journal
This document discusses performance analysis of fingerprint image compression using two techniques: K-SVD-SR and SPIHT. K-SVD-SR is a novel compression algorithm based on K-singular value decomposition and sparse representation that has the ability to update the dictionary. The document compares the compressed image quality of K-SVD-SR to SPIHT in terms of mean square error and peak signal-to-noise ratio. It also describes the methodology used in K-SVD-SR, including constructing the dictionary from a training set of fingerprint patches and compressing new fingerprints by representing patches as sparse combinations of dictionary atoms.
The document summarizes an algorithm for image encryption and compression based on compressive sensing and chaos. It begins with background information on compressive sensing theory and multi-chaotic based image encryption. It then describes the proposed algorithm which uses compressive sensing and a multi-chaotic system together for both image encryption and compression in a single step. Simulation results showed that the encrypted images had a large key space, low storage and transmission requirements, high security, and good statistical properties. Recovered images also had good quality while preserving image characteristics.
Importance of Dimensionality Reduction in Image Processingrahulmonikasharma
This paper presents a survey on various techniques of compression methods. Linear Discriminant analysis (LDA) is a method used in statistics, pattern recognition and machine learning to find a linear combination of features that classifies an object into two or more classes. This results in a dimensionality reduction before later classification.Principal component analysis (PCA) uses an orthogonal transformation to convert a set of correlated variables into a set of values of linearly uncorrelated variables called principal components. The purpose of the review is to explore the possibility of image compression for multiple images.
Compression technique using dct fractal compressionAlexander Decker
This document summarizes and compares different image compression techniques, including DCT, fractal compression, and their applications in steganography. It discusses how DCT works by transforming image data into frequency domains, while fractal compression exploits self-similarity within images. The document reviews several existing studies on combining these techniques with steganography and encryption. Specifically, it examines approaches that use DCT and fractal compression to improve data hiding capacity and security. Overall, the document provides an overview of key compression algorithms and their applications in digital watermarking and steganography.
11.compression technique using dct fractal compressionAlexander Decker
1) The document discusses and compares different image compression techniques, specifically DCT and fractal compression.
2) Fractal compression works by finding self-similar patterns within an image during encoding, but can have a long computation time. DCT transforms an image into frequency coefficients that can be quantized for compression.
3) The document reviews previous work combining DCT and fractal compression with steganography and encryption to improve hiding capacity, imperceptibility, and security against subterfuge attacks. However, prior methods had limitations like low data hiding amounts or lack of protection for compressed data.
A common goal of the engineering field of signal processing is to reconstruct a signal from a series of sampling measurements. In general, this task is impossible because there is no way to reconstruct a signal during the times
that the signal is not measured. Nevertheless, with prior knowledge or assumptions about the signal, it turns out to
be possible to perfectly reconstruct a signal from a series of measurements. Over time, engineers have improved their understanding of which assumptions are practical and how they can be generalized. An early breakthrough in signal processing was the Nyquist–Shannon sampling theorem. It states that if the signal's highest frequency is less than half of the sampling rate, then the signal can be reconstructed perfectly. The main idea is that with prior knowledge about constraints on the signal’s frequencies, fewer samples are needed to reconstruct the signal. Sparse sampling (also known as, compressive sampling, or compressed sampling) is a signal processing technique for efficiently acquiring and reconstructing a signal, by finding solutions tounder determined linear systems. This is based on the principle that, through optimization, the sparsity of a signal can be exploited to recover it from far fewer samples than required by the Shannon-Nyquist sampling theorem. There are two conditions under which recovery is possible.[1] The first one is sparsity which requires the signal to be sparse in some domain. The second one is incoherence which is applied through the isometric property which is sufficient for sparse signals Possibility
of compressed data acquisition protocols which directly acquire just the important information Sparse sampling (CS) is a fast growing area of research. It neglects the extravagant acquisition process by measuring lesser values to reconstruct the image or signal. Sparse sampling is adopted successfully in various fields of image processing and proved its efficiency. Some of the image processing applications like face recognition, video encoding, Image encryption and reconstruction are presented here.
Provably secure and efficient audio compression based on compressive sensingIJECEIAES
The advancement of systems with the capacity to compress audio signals and simultaneously secure is a highly attractive research subject. This is because of the need to enhance storage usage and speed up the transmission of data, as well as securing the transmission of sensitive signals over limited and insecure communication channels. Thus, many researchers have studied and produced different systems, either to compress or encrypt audio data using different algorithms and methods, all of which suffer from certain issues including high time consumption or complex calculations. This paper proposes a compressing sensing-based system that compresses audio signals and simultaneously provides an encryption system. The audio signal is segmented into small matrices of samples and then multiplied by a non-square sensing matrix generated by a Gaussian random generator. The reconstruction process is carried out by solving a linear system using the pseudoinverse of Moore-Penrose. The statistical analysis results obtaining from implementing different types and sizes of audio signals prove that the proposed system succeeds in compressing the audio signals with a ratio reaching 28% of real size and reconstructing the signal with a correlation metric between 0.98 and 0.99. It also scores very good results in the normalized mean square error (MSE), peak signal-to-noise ratio metrics (PSNR), and the structural similarity index (SSIM), as well as giving the signal a high level of security.
Redefining brain tumor segmentation: a cutting-edge convolutional neural netw...IJECEIAES
Medical image analysis has witnessed significant advancements with deep learning techniques. In the domain of brain tumor segmentation, the ability to
precisely delineate tumor boundaries from magnetic resonance imaging (MRI)
scans holds profound implications for diagnosis. This study presents an ensemble convolutional neural network (CNN) with transfer learning, integrating
the state-of-the-art Deeplabv3+ architecture with the ResNet18 backbone. The
model is rigorously trained and evaluated, exhibiting remarkable performance
metrics, including an impressive global accuracy of 99.286%, a high-class accuracy of 82.191%, a mean intersection over union (IoU) of 79.900%, a weighted
IoU of 98.620%, and a Boundary F1 (BF) score of 83.303%. Notably, a detailed comparative analysis with existing methods showcases the superiority of
our proposed model. These findings underscore the model’s competence in precise brain tumor localization, underscoring its potential to revolutionize medical
image analysis and enhance healthcare outcomes. This research paves the way
for future exploration and optimization of advanced CNN models in medical
imaging, emphasizing addressing false positives and resource efficiency.
Embedded machine learning-based road conditions and driving behavior monitoringIJECEIAES
Car accident rates have increased in recent years, resulting in losses in human lives, properties, and other financial costs. An embedded machine learning-based system is developed to address this critical issue. The system can monitor road conditions, detect driving patterns, and identify aggressive driving behaviors. The system is based on neural networks trained on a comprehensive dataset of driving events, driving styles, and road conditions. The system effectively detects potential risks and helps mitigate the frequency and impact of accidents. The primary goal is to ensure the safety of drivers and vehicles. Collecting data involved gathering information on three key road events: normal street and normal drive, speed bumps, circular yellow speed bumps, and three aggressive driving actions: sudden start, sudden stop, and sudden entry. The gathered data is processed and analyzed using a machine learning system designed for limited power and memory devices. The developed system resulted in 91.9% accuracy, 93.6% precision, and 92% recall. The achieved inference time on an Arduino Nano 33 BLE Sense with a 32-bit CPU running at 64 MHz is 34 ms and requires 2.6 kB peak RAM and 139.9 kB program flash memory, making it suitable for resource-constrained embedded systems.
Advanced control scheme of doubly fed induction generator for wind turbine us...IJECEIAES
This paper describes a speed control device for generating electrical energy on an electricity network based on the doubly fed induction generator (DFIG) used for wind power conversion systems. At first, a double-fed induction generator model was constructed. A control law is formulated to govern the flow of energy between the stator of a DFIG and the energy network using three types of controllers: proportional integral (PI), sliding mode controller (SMC) and second order sliding mode controller (SOSMC). Their different results in terms of power reference tracking, reaction to unexpected speed fluctuations, sensitivity to perturbations, and resilience against machine parameter alterations are compared. MATLAB/Simulink was used to conduct the simulations for the preceding study. Multiple simulations have shown very satisfying results, and the investigations demonstrate the efficacy and power-enhancing capabilities of the suggested control system.
Neural network optimizer of proportional-integral-differential controller par...IJECEIAES
Wide application of proportional-integral-differential (PID)-regulator in industry requires constant improvement of methods of its parameters adjustment. The paper deals with the issues of optimization of PID-regulator parameters with the use of neural network technology methods. A methodology for choosing the architecture (structure) of neural network optimizer is proposed, which consists in determining the number of layers, the number of neurons in each layer, as well as the form and type of activation function. Algorithms of neural network training based on the application of the method of minimizing the mismatch between the regulated value and the target value are developed. The method of back propagation of gradients is proposed to select the optimal training rate of neurons of the neural network. The neural network optimizer, which is a superstructure of the linear PID controller, allows increasing the regulation accuracy from 0.23 to 0.09, thus reducing the power consumption from 65% to 53%. The results of the conducted experiments allow us to conclude that the created neural superstructure may well become a prototype of an automatic voltage regulator (AVR)-type industrial controller for tuning the parameters of the PID controller.
An improved modulation technique suitable for a three level flying capacitor ...IJECEIAES
This research paper introduces an innovative modulation technique for controlling a 3-level flying capacitor multilevel inverter (FCMLI), aiming to streamline the modulation process in contrast to conventional methods. The proposed
simplified modulation technique paves the way for more straightforward and
efficient control of multilevel inverters, enabling their widespread adoption and
integration into modern power electronic systems. Through the amalgamation of
sinusoidal pulse width modulation (SPWM) with a high-frequency square wave
pulse, this controlling technique attains energy equilibrium across the coupling
capacitor. The modulation scheme incorporates a simplified switching pattern
and a decreased count of voltage references, thereby simplifying the control
algorithm.
A review on features and methods of potential fishing zoneIJECEIAES
This review focuses on the importance of identifying potential fishing zones in seawater for sustainable fishing practices. It explores features like sea surface temperature (SST) and sea surface height (SSH), along with classification methods such as classifiers. The features like SST, SSH, and different classifiers used to classify the data, have been figured out in this review study. This study underscores the importance of examining potential fishing zones using advanced analytical techniques. It thoroughly explores the methodologies employed by researchers, covering both past and current approaches. The examination centers on data characteristics and the application of classification algorithms for classification of potential fishing zones. Furthermore, the prediction of potential fishing zones relies significantly on the effectiveness of classification algorithms. Previous research has assessed the performance of models like support vector machines, naïve Bayes, and artificial neural networks (ANN). In the previous result, the results of support vector machine (SVM) were 97.6% more accurate than naive Bayes's 94.2% to classify test data for fisheries classification. By considering the recent works in this area, several recommendations for future works are presented to further improve the performance of the potential fishing zone models, which is important to the fisheries community.
Electrical signal interference minimization using appropriate core material f...IJECEIAES
As demand for smaller, quicker, and more powerful devices rises, Moore's law is strictly followed. The industry has worked hard to make little devices that boost productivity. The goal is to optimize device density. Scientists are reducing connection delays to improve circuit performance. This helped them understand three-dimensional integrated circuit (3D IC) concepts, which stack active devices and create vertical connections to diminish latency and lower interconnects. Electrical involvement is a big worry with 3D integrates circuits. Researchers have developed and tested through silicon via (TSV) and substrates to decrease electrical wave involvement. This study illustrates a novel noise coupling reduction method using several electrical involvement models. A 22% drop in electrical involvement from wave-carrying to victim TSVs introduces this new paradigm and improves system performance even at higher THz frequencies.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
Bibliometric analysis highlighting the role of women in addressing climate ch...IJECEIAES
Fossil fuel consumption increased quickly, contributing to climate change
that is evident in unusual flooding and draughts, and global warming. Over
the past ten years, women's involvement in society has grown dramatically,
and they succeeded in playing a noticeable role in reducing climate change.
A bibliometric analysis of data from the last ten years has been carried out to
examine the role of women in addressing the climate change. The analysis's
findings discussed the relevant to the sustainable development goals (SDGs),
particularly SDG 7 and SDG 13. The results considered contributions made
by women in the various sectors while taking geographic dispersion into
account. The bibliometric analysis delves into topics including women's
leadership in environmental groups, their involvement in policymaking, their
contributions to sustainable development projects, and the influence of
gender diversity on attempts to mitigate climate change. This study's results
highlight how women have influenced policies and actions related to climate
change, point out areas of research deficiency and recommendations on how
to increase role of the women in addressing the climate change and
achieving sustainability. To achieve more successful results, this initiative
aims to highlight the significance of gender equality and encourage
inclusivity in climate change decision-making processes.
Voltage and frequency control of microgrid in presence of micro-turbine inter...IJECEIAES
The active and reactive load changes have a significant impact on voltage
and frequency. In this paper, in order to stabilize the microgrid (MG) against
load variations in islanding mode, the active and reactive power of all
distributed generators (DGs), including energy storage (battery), diesel
generator, and micro-turbine, are controlled. The micro-turbine generator is
connected to MG through a three-phase to three-phase matrix converter, and
the droop control method is applied for controlling the voltage and
frequency of MG. In addition, a method is introduced for voltage and
frequency control of micro-turbines in the transition state from gridconnected mode to islanding mode. A novel switching strategy of the matrix
converter is used for converting the high-frequency output voltage of the
micro-turbine to the grid-side frequency of the utility system. Moreover,
using the switching strategy, the low-order harmonics in the output current
and voltage are not produced, and consequently, the size of the output filter
would be reduced. In fact, the suggested control strategy is load-independent
and has no frequency conversion restrictions. The proposed approach for
voltage and frequency regulation demonstrates exceptional performance and
favorable response across various load alteration scenarios. The suggested
strategy is examined in several scenarios in the MG test systems, and the
simulation results are addressed.
Enhancing battery system identification: nonlinear autoregressive modeling fo...IJECEIAES
Precisely characterizing Li-ion batteries is essential for optimizing their
performance, enhancing safety, and prolonging their lifespan across various
applications, such as electric vehicles and renewable energy systems. This
article introduces an innovative nonlinear methodology for system
identification of a Li-ion battery, employing a nonlinear autoregressive with
exogenous inputs (NARX) model. The proposed approach integrates the
benefits of nonlinear modeling with the adaptability of the NARX structure,
facilitating a more comprehensive representation of the intricate
electrochemical processes within the battery. Experimental data collected
from a Li-ion battery operating under diverse scenarios are employed to
validate the effectiveness of the proposed methodology. The identified
NARX model exhibits superior accuracy in predicting the battery's behavior
compared to traditional linear models. This study underscores the
importance of accounting for nonlinearities in battery modeling, providing
insights into the intricate relationships between state-of-charge, voltage, and
current under dynamic conditions.
Smart grid deployment: from a bibliometric analysis to a surveyIJECEIAES
Smart grids are one of the last decades' innovations in electrical energy.
They bring relevant advantages compared to the traditional grid and
significant interest from the research community. Assessing the field's
evolution is essential to propose guidelines for facing new and future smart
grid challenges. In addition, knowing the main technologies involved in the
deployment of smart grids (SGs) is important to highlight possible
shortcomings that can be mitigated by developing new tools. This paper
contributes to the research trends mentioned above by focusing on two
objectives. First, a bibliometric analysis is presented to give an overview of
the current research level about smart grid deployment. Second, a survey of
the main technological approaches used for smart grid implementation and
their contributions are highlighted. To that effect, we searched the Web of
Science (WoS), and the Scopus databases. We obtained 5,663 documents
from WoS and 7,215 from Scopus on smart grid implementation or
deployment. With the extraction limitation in the Scopus database, 5,872 of
the 7,215 documents were extracted using a multi-step process. These two
datasets have been analyzed using a bibliometric tool called bibliometrix.
The main outputs are presented with some recommendations for future
research.
Use of analytical hierarchy process for selecting and prioritizing islanding ...IJECEIAES
One of the problems that are associated to power systems is islanding
condition, which must be rapidly and properly detected to prevent any
negative consequences on the system's protection, stability, and security.
This paper offers a thorough overview of several islanding detection
strategies, which are divided into two categories: classic approaches,
including local and remote approaches, and modern techniques, including
techniques based on signal processing and computational intelligence.
Additionally, each approach is compared and assessed based on several
factors, including implementation costs, non-detected zones, declining
power quality, and response times using the analytical hierarchy process
(AHP). The multi-criteria decision-making analysis shows that the overall
weight of passive methods (24.7%), active methods (7.8%), hybrid methods
(5.6%), remote methods (14.5%), signal processing-based methods (26.6%),
and computational intelligent-based methods (20.8%) based on the
comparison of all criteria together. Thus, it can be seen from the total weight
that hybrid approaches are the least suitable to be chosen, while signal
processing-based methods are the most appropriate islanding detection
method to be selected and implemented in power system with respect to the
aforementioned factors. Using Expert Choice software, the proposed
hierarchy model is studied and examined.
Enhancing of single-stage grid-connected photovoltaic system using fuzzy logi...IJECEIAES
The power generated by photovoltaic (PV) systems is influenced by
environmental factors. This variability hampers the control and utilization of
solar cells' peak output. In this study, a single-stage grid-connected PV
system is designed to enhance power quality. Our approach employs fuzzy
logic in the direct power control (DPC) of a three-phase voltage source
inverter (VSI), enabling seamless integration of the PV connected to the
grid. Additionally, a fuzzy logic-based maximum power point tracking
(MPPT) controller is adopted, which outperforms traditional methods like
incremental conductance (INC) in enhancing solar cell efficiency and
minimizing the response time. Moreover, the inverter's real-time active and
reactive power is directly managed to achieve a unity power factor (UPF).
The system's performance is assessed through MATLAB/Simulink
implementation, showing marked improvement over conventional methods,
particularly in steady-state and varying weather conditions. For solar
irradiances of 500 and 1,000 W/m2
, the results show that the proposed
method reduces the total harmonic distortion (THD) of the injected current
to the grid by approximately 46% and 38% compared to conventional
methods, respectively. Furthermore, we compare the simulation results with
IEEE standards to evaluate the system's grid compatibility.
Enhancing photovoltaic system maximum power point tracking with fuzzy logic-b...IJECEIAES
Photovoltaic systems have emerged as a promising energy resource that
caters to the future needs of society, owing to their renewable, inexhaustible,
and cost-free nature. The power output of these systems relies on solar cell
radiation and temperature. In order to mitigate the dependence on
atmospheric conditions and enhance power tracking, a conventional
approach has been improved by integrating various methods. To optimize
the generation of electricity from solar systems, the maximum power point
tracking (MPPT) technique is employed. To overcome limitations such as
steady-state voltage oscillations and improve transient response, two
traditional MPPT methods, namely fuzzy logic controller (FLC) and perturb
and observe (P&O), have been modified. This research paper aims to
simulate and validate the step size of the proposed modified P&O and FLC
techniques within the MPPT algorithm using MATLAB/Simulink for
efficient power tracking in photovoltaic systems.
Adaptive synchronous sliding control for a robot manipulator based on neural ...IJECEIAES
Robot manipulators have become important equipment in production lines, medical fields, and transportation. Improving the quality of trajectory tracking for
robot hands is always an attractive topic in the research community. This is a
challenging problem because robot manipulators are complex nonlinear systems
and are often subject to fluctuations in loads and external disturbances. This
article proposes an adaptive synchronous sliding control scheme to improve trajectory tracking performance for a robot manipulator. The proposed controller
ensures that the positions of the joints track the desired trajectory, synchronize
the errors, and significantly reduces chattering. First, the synchronous tracking
errors and synchronous sliding surfaces are presented. Second, the synchronous
tracking error dynamics are determined. Third, a robust adaptive control law is
designed,the unknown components of the model are estimated online by the neural network, and the parameters of the switching elements are selected by fuzzy
logic. The built algorithm ensures that the tracking and approximation errors
are ultimately uniformly bounded (UUB). Finally, the effectiveness of the constructed algorithm is demonstrated through simulation and experimental results.
Simulation and experimental results show that the proposed controller is effective with small synchronous tracking errors, and the chattering phenomenon is
significantly reduced.
Remote field-programmable gate array laboratory for signal acquisition and de...IJECEIAES
A remote laboratory utilizing field-programmable gate array (FPGA) technologies enhances students’ learning experience anywhere and anytime in embedded system design. Existing remote laboratories prioritize hardware access and visual feedback for observing board behavior after programming, neglecting comprehensive debugging tools to resolve errors that require internal signal acquisition. This paper proposes a novel remote embeddedsystem design approach targeting FPGA technologies that are fully interactive via a web-based platform. Our solution provides FPGA board access and debugging capabilities beyond the visual feedback provided by existing remote laboratories. We implemented a lab module that allows users to seamlessly incorporate into their FPGA design. The module minimizes hardware resource utilization while enabling the acquisition of a large number of data samples from the signal during the experiments by adaptively compressing the signal prior to data transmission. The results demonstrate an average compression ratio of 2.90 across three benchmark signals, indicating efficient signal acquisition and effective debugging and analysis. This method allows users to acquire more data samples than conventional methods. The proposed lab allows students to remotely test and debug their designs, bridging the gap between theory and practice in embedded system design.
Detecting and resolving feature envy through automated machine learning and m...IJECEIAES
Efficiently identifying and resolving code smells enhances software project quality. This paper presents a novel solution, utilizing automated machine learning (AutoML) techniques, to detect code smells and apply move method refactoring. By evaluating code metrics before and after refactoring, we assessed its impact on coupling, complexity, and cohesion. Key contributions of this research include a unique dataset for code smell classification and the development of models using AutoGluon for optimal performance. Furthermore, the study identifies the top 20 influential features in classifying feature envy, a well-known code smell, stemming from excessive reliance on external classes. We also explored how move method refactoring addresses feature envy, revealing reduced coupling and complexity, and improved cohesion, ultimately enhancing code quality. In summary, this research offers an empirical, data-driven approach, integrating AutoML and move method refactoring to optimize software project quality. Insights gained shed light on the benefits of refactoring on code quality and the significance of specific features in detecting feature envy. Future research can expand to explore additional refactoring techniques and a broader range of code metrics, advancing software engineering practices and standards.
Smart monitoring technique for solar cell systems using internet of things ba...IJECEIAES
Rapidly and remotely monitoring and receiving the solar cell systems status parameters, solar irradiance, temperature, and humidity, are critical issues in enhancement their efficiency. Hence, in the present article an improved smart prototype of internet of things (IoT) technique based on embedded system through NodeMCU ESP8266 (ESP-12E) was carried out experimentally. Three different regions at Egypt; Luxor, Cairo, and El-Beheira cities were chosen to study their solar irradiance profile, temperature, and humidity by the proposed IoT system. The monitoring data of solar irradiance, temperature, and humidity were live visualized directly by Ubidots through hypertext transfer protocol (HTTP) protocol. The measured solar power radiation in Luxor, Cairo, and El-Beheira ranged between 216-1000, 245-958, and 187-692 W/m 2 respectively during the solar day. The accuracy and rapidity of obtaining monitoring results using the proposed IoT system made it a strong candidate for application in monitoring solar cell systems. On the other hand, the obtained solar power radiation results of the three considered regions strongly candidate Luxor and Cairo as suitable places to build up a solar cells system station rather than El-Beheira.
An efficient security framework for intrusion detection and prevention in int...IJECEIAES
Over the past few years, the internet of things (IoT) has advanced to connect billions of smart devices to improve quality of life. However, anomalies or malicious intrusions pose several security loopholes, leading to performance degradation and threat to data security in IoT operations. Thereby, IoT security systems must keep an eye on and restrict unwanted events from occurring in the IoT network. Recently, various technical solutions based on machine learning (ML) models have been derived towards identifying and restricting unwanted events in IoT. However, most ML-based approaches are prone to miss-classification due to inappropriate feature selection. Additionally, most ML approaches applied to intrusion detection and prevention consider supervised learning, which requires a large amount of labeled data to be trained. Consequently, such complex datasets are impossible to source in a large network like IoT. To address this problem, this proposed study introduces an efficient learning mechanism to strengthen the IoT security aspects. The proposed algorithm incorporates supervised and unsupervised approaches to improve the learning models for intrusion detection and mitigation. Compared with the related works, the experimental outcome shows that the model performs well in a benchmark dataset. It accomplishes an improved detection accuracy of approximately 99.21%.
Main Menu The metals-black-book-ferrous-metalsRicardo Akerman
Guia técnico e de referência amplamente utilizado nas indústrias metalúrgica, de manufatura, engenharia, petróleo e gás, construção naval, e diversas áreas de manutenção industrial.
MODULE 5 BUILDING PLANNING AND DESIGN SY BTECH ACOUSTICS SYSTEM IN BUILDINGDr. BASWESHWAR JIRWANKAR
: Introduction to Acoustics & Green Building -
Absorption of sound, various materials, Sabine’s formula, optimum reverberation time, conditions for good acoustics Sound insulation:
Acceptable noise levels, noise prevention at its source, transmission of noise, Noise control-general considerations
Green Building: Concept, Principles, Materials, Characteristics, Applications
ISO 4020-6.1 – Filter Cleanliness Test Rig: Precision Testing for Fuel Filter Integrity
Explore the design, functionality, and standards compliance of our advanced Filter Cleanliness Test Rig developed according to ISO 4020-6.1. This rig is engineered to evaluate fuel filter cleanliness levels with high accuracy and repeatability—critical for ensuring the performance and durability of fuel systems.
🔬 Inside This Presentation:
Overview of ISO 4020-6.1 testing protocols
Rig components and schematic layout
Test methodology and data acquisition
Applications in automotive and industrial filtration
Key benefits: accuracy, reliability, compliance
Perfect for R&D engineers, quality assurance teams, and lab technicians focused on filtration performance and standard compliance.
🛠️ Ensure Filter Cleanliness — Validate with Confidence.
This presentation provides a comprehensive overview of a specialized test rig designed in accordance with ISO 4548-7, the international standard for evaluating the vibration fatigue resistance of full-flow lubricating oil filters used in internal combustion engines.
Key features include:
Filters for Electromagnetic Compatibility ApplicationsMathias Magdowski
In this lecture, I explain the fundamentals of electromagnetic compatibility (EMC), the basic coupling model and coupling paths via cables, electric fields, magnetic fields and wave fields. We also look at electric vehicles as an example of systems with many conducted EMC problems due to power electronic devices such as rectifiers and inverters with non-linear components such as diodes and fast switching components such as MOSFETs or IGBTs. After a brief review of circuit analysis fundamentals and an experimental investigation of the frequency-dependent impedance of resistors, capacitors and inductors, we look at a simple low-pass filter. The input impedance from both sides as well as the transfer function are measured.
Department of Environment (DOE) Mix Design with Fly Ash.MdManikurRahman
Concrete Mix Design with Fly Ash by DOE Method. The Department of Environmental (DOE) approach to fly ash-based concrete mix design is covered in this study.
The Department of Environment (DOE) method of mix design is a British method originally developed in the UK in the 1970s. It is widely used for concrete mix design, including mixes that incorporate supplementary cementitious materials (SCMs) such as fly ash.
When using fly ash in concrete, the DOE method can be adapted to account for its properties and effects on workability, strength, and durability. Here's a step-by-step overview of how the DOE method is applied with fly ash.
Peak ground acceleration (PGA) is a critical parameter in ground-motion investigations, in particular in earthquake-prone areas such as Iran. In the current study, a new method based on particle swarm optimization (PSO) is developed to obtain an efficient attenuation relationship for the vertical PGA component within the northern Iranian plateau. The main purpose of this study is to propose suitable attenuation relationships for calculating the PGA for the Alborz, Tabriz and Kopet Dag faults in the vertical direction. To this aim, the available catalogs of the study area are investigated, and finally about 240 earthquake records (with a moment magnitude of 4.1 to 6.4) are chosen to develop the model. Afterward, the PSO algorithm is used to estimate model parameters, i.e., unknown coefficients of the model (attenuation relationship). Different statistical criteria showed the acceptable performance of the proposed relationships in the estimation of vertical PGA components in comparison to the previously developed relationships for the northern plateau of Iran. Developed attenuation relationships in the current study are independent of shear wave velocity. This issue is the advantage of proposed relationships for utilizing in the situations where there are not sufficient shear wave velocity data.
UNIT-1-PPT-Introduction about Power System Operation and ControlSridhar191373
Power scenario in Indian grid – National and Regional load dispatching centers –requirements of good power system - necessity of voltage and frequency regulation – real power vs frequency and reactive power vs voltage control loops - system load variation, load curves and basic concepts of load dispatching - load forecasting - Basics of speed governing mechanisms and modeling - speed load characteristics - regulation of two generators in parallel.
Expansive soils (ES) have a long history of being difficult to work with in geotechnical engineering. Numerous studies have examined how bagasse ash (BA) and lime affect the unconfined compressive strength (UCS) of ES. Due to the complexities of this composite material, determining the UCS of stabilized ES using traditional methods such as empirical approaches and experimental methods is challenging. The use of artificial neural networks (ANN) for forecasting the UCS of stabilized soil has, however, been the subject of a few studies. This paper presents the results of using rigorous modelling techniques like ANN and multi-variable regression model (MVR) to examine the UCS of BA and a blend of BA-lime (BA + lime) stabilized ES. Laboratory tests were conducted for all dosages of BA and BA-lime admixed ES. 79 samples of data were gathered with various combinations of the experimental variables prepared and used in the construction of ANN and MVR models. The input variables for two models are seven parameters: BA percentage, lime percentage, liquid limit (LL), plastic limit (PL), shrinkage limit (SL), maximum dry density (MDD), and optimum moisture content (OMC), with the output variable being 28-day UCS. The ANN model prediction performance was compared to that of the MVR model. The models were evaluated and contrasted on the training dataset (70% data) and the testing dataset (30% residual data) using the coefficient of determination (R2), Mean Absolute Error (MAE), and Root Mean Square Error (RMSE) criteria. The findings indicate that the ANN model can predict the UCS of stabilized ES with high accuracy. The relevance of various input factors was estimated via sensitivity analysis utilizing various methodologies. For both the training and testing data sets, the proposed model has an elevated R2 of 0.9999. It has a minimal MAE and RMSE value of 0.0042 and 0.0217 for training data and 0.0038 and 0.0104 for testing data. As a result, the generated model excels the MVR model in terms of UCS prediction.
This presentation provides a detailed overview of air filter testing equipment, including its types, working principles, and industrial applications. Learn about key performance indicators such as filtration efficiency, pressure drop, and particulate holding capacity. The slides highlight standard testing methods (e.g., ISO 16890, EN 1822, ASHRAE 52.2), equipment configurations (such as aerosol generators, particle counters, and test ducts), and the role of automation and data logging in modern systems. Ideal for engineers, quality assurance professionals, and researchers involved in HVAC, automotive, cleanroom, or industrial filtration systems.
2. IJECE ISSN: 2088-8708
Dictionary based Image Compression via Sparse Representation (Arabinda Sahoo)
1965
In contrast to fixed DCT and wavelet dictionary, the latest trend of image compression techniques is
extended to design trained dictionaries [12]. Numerous dictionary based image compression methods have
been proposed for sparse representation. In several recently published works, the use of learned over-
complete dictionaries in image compression has shown promising results at low bit rate as compared to fixed
dictionaries. The initial proposal towards dictionary based image compression was proposed by Bryt and
Elad [13]. They proposed an algorithm for the compression of facial image based on the learned K-SVD [14]
over-complete dictionary. Though this method outperforms JPEG and JPEG2000 but this method is limited
to compressing facial images. In [15], the author proposed an image compression method based on the
iteration-tuned dictionary (ITD). In this scheme, the dictionary consists of a layer structure with each layer
trained for a specific class of images and carries a separate dictionary matrix. This method is shown to
outperform K-SVD over-complete dictionary method but it is employed to compress a specific class of
images. In [16], the author proposed an adaptive dictionary design method for the fingerprint image
compression. They used a set of fingerprint image to learn a dictionary. In [17], the author proposed a
dictionary based method for compressing surveillance image.
Experimental result shows these above proposed algorithms outperform JPEG and JPEG2000.
However, most of the above mentioned compression scheme is either related to facial images or to some
specific class of images, while there is a lack of research on general arbitrary images. So, the main challenge
of above schemes is the compression of general arbitrary images. To address this, in this paper we proposed a
novel image compression scheme for arbitrary images. In this scheme, an efficient trained over-complete
dictionary is integrated into the intra-prediction framework. The conventional transform-domain
representation of intra-prediction scheme is replaced by a trained over-complete dictionary. We trained a
dictionary offline using the residuals obtained from intra-prediction. This dictionary can sparsely represent
the complex characteristics of the residual block. The coefficients and indices of appropriate dictionary
element obtained from sparse representation are transmitted for encoding. Since the dictionary is shared at
both encoder and decoder, only coefficients and indices of dictionary elements need to be encoded, which
compress the image significantly. Experimental results on the arbitrary images shows that the proposed
method yields both improved coding efficiency and image quality as compared to JPEG and JPEG 2000.
The rest of the paper is organized as follows. In Section 2, we present some preliminaries on sparse
representation and dictionary design. The proposed image compression method is introduced in section 3.
Section 4 illustrates experimental results and discussion. Finally, Section 5 concludes the paper.
2. PRELIMINARIES
In sparse representation [18], a signal can be represented by a linear combination of a small
number of signals known as atoms taken from an over-complete dictionary . It is called sparse
representation as it employs only a few number dictionary atoms or elements to represent the signal. A signal
is said to be compressible if it can be represented by few dictionary atoms. Mathematically sparse
representation can be expressed as:
(1)
The solution vector contains the coefficients of the signal . Where D is one n×k matrix with n < k
called over-complete dictionary and each column of D is called an atom. If and each atom of D are treated
as a signal then can be represented as a linear combination of atoms of D. This linear combination can be
expressed as solution vector . Due to over complete nature of D, an infinite number of solution exist for .
In last decade, various sparse approximation algorithms have been proposed to find out the sparse solution
for . The sparse approximation algorithm always aims to represent in terms of minimum number atoms.
Mathematically, this can be expressed as solving Equation (1) such that the solution contains minimum
number of non-zero elements. A signal is said to be compressible if the number of non-zero elements in is
very less as compared to number of elements of . The sparse representation of may be either exact
or approximate, satisfying .Where p is the norm.Typical norms used in the
approximation are 1, 2 or ∞. Normally p = 2 is taken in image compression. Mathematically, the sprarsest
representation is the solution of either Equations (2) or (3)
(2)
or
(3)
3. ISSN: 2088-8708
IJECE Vol. 7, No. 4, August 2017 : 1964 – 1972
1966
where is the l0 norm, represents the number of non-zero elements in solution vector x. Some of the well
known algorithms used to find sparse representation are: Matching pursuit (MP) [19], Orthogonal Matching
Pursuit (OMP) [20] and Complimentary matching pursuit (CMP) [21]. In this paper, we focus on OMP
algorithm due to its efficiency.
The dictionary based image compression can also be effectively modeled by Equations (2) and (3).
In image compression, we consider a set of k image blocks of size n pixels with n < k, ordered
lexicographically as column vectors of dictionary D. A column vector y is obtained from an image block of
size n pixels. In sparse representation, the problem is to find out the solution vector which will represent y
with least number of dictionary elements. Indeed, compression of image patch y can be achieved by
transmission of nonzero elements of vector , by specifying their coefficients and indices.
The dictionary plays an important role in a successful image compression modeling via sparse
representation. Am image is compressible if it can be represented by few number of dictionary elements. The
dictionary can either be chosen as a prespecified set of images or designed by adapting its contents to fit a
given set of images. The objective of dictionary design is to train the dictionary which able to represent a
signal set sparsely [12]. Given an image set { } , dictionary design aims to find the best dictionary D
that gives rise to sparse solution for each . In other words, there exists D, such that solving Equation (2)
for each gives a sparse representation . The minimization problem to find the best dictionary for sparse
representation of Y in the given sparsity constraint T0 can be represented by:
(4)
The dictionary is trained to provide a better representation of the actual signal when the number of
dictionary elements used to represent it is less than or equal to T0.Various algorithms have been developed to
train over complete dictionaries for sparse signal representation. The K-SVD algorithm [14] is very efficient
and it works well with different sparse approximation algorithm. K-SVD algorithm iteratively updates the
dictionary atoms to better fit the data. In this paper, we focus on K-SVD algorithm to train the dictionary and
OMP algorithm for sparse representation.
(a) Encoder
(b) Decoder
Figure 1. Detailed block diagram of the proposed method
4. IJECE ISSN: 2088-8708
Dictionary based Image Compression via Sparse Representation (Arabinda Sahoo)
1967
3. PROPOSED METHOD
The proposed image compression method consists of four main processes: intra prediction,
dictionary training, sparse representation, and coding. The block diagram of the proposed compression
method is shown in Figure 1.
3.1. Intra Prediction
Intra prediction [8] exploits spatial correlation within one image. In this process, the current block is
predicted by using the boundary pixel values of previously reconstructed neighboring blocks. As shown in
the Figure 2, the boundary pixel values of neighboring blocks as shown in shaded boxes are copied into the
current block pixels along a specified direction indicated by the mode. Eight directional modes and a DC
mode, which is almost same as the nine intra-prediction modes employed in H.264 standard [9].
(a) (b)
Figure 2. (a) Intra Prediction, (b) 8-Directional mode + DC Mode
In this proposed method, the image is divided into blocks of size 8 x 8. The nine intra-prediction
modes (0-8) are applied over each 8 x 8 block and residual error is calculated for each prediction mode. The
residual error is the difference between the pixel value of the current block and pixel value of the predicted
block. The best prediction mode for a current block is selected based on the minimum residual error. The
mode number M and the residual error block are transmitted for encoding.
3.2. Dictionary Training
The dictionary is trained off-line using prediction residual samples resulting from a wide variety of
images. We first divide different images into blocks of size 8 x 8 and then 9 intra- prediction modes (0-8) are
applied over each 8 x 8 block. The predicted block is subtracted from the current block to generate residual
blocks. A set of 8 × 8 prediction residual blocks for different modes are selected to train the dictionary. To
train the dictionary we employed K-SVD algorithm [14]. During dictionary training, in each iteration, K-
SVD algorithm updates the dictionary elements by optimizing minimization problem given in Equation(4).
K-SVD algorithm iteratively updates the dictionary elements to better fit the data and after certain iteration,
an updated dictionary is resulted. This updated dictionary is used for sparse representation of the residual
block during image coding.
3.3. Sparse Representation
OMP algorithm is employed for the sparse representation of residual block. OMP algorithm selects
the appropriate dictionary element to represent each residual image block. In each iteration, OMP selects the
best linear combination of dictionary elements by minimizing Equation (3). The same process is continued
and the algorithm terminates when the residual error of the reconstructed signal is equal to or less than a
specified value. However, the number of OMP iterations may not exceed T0. Once the algorithm terminates
the coefficients C and indices P of appropriate dictionary element is transmitted for encoding.
3.4. Coding
After sparse representation, the coefficients C and indices P of dictionary element, and the
prediction mode M are encoded [22]. The coefficients are uniformly quantized followed by entropy coding.
5. ISSN: 2088-8708
IJECE Vol. 7, No. 4, August 2017 : 1964 – 1972
1968
The indices are encoded with fixed length codes whose sizes are log2k, where k is the number of dictionary
elements. The prediction mode numbers are encoded with fixed length codes whose sizes are 4 bits. The
encoder of the proposed compression method is shown in Figure 1(a).
Since the dictionary is shared at both encoder and decoder, only intra- prediction mode number,
indices, and coefficients of the dictionary elements are transmitted to the decoder. The decoder generates the
residual block from the knowledge of dictionary, coefficients, and indices. The decoder then predicts the
block based on mode number and combines with the residual block to reconstruct the block. The decoder of
the proposed compression method is shown in Figure 1(b).
4. EXPERIMENTAL RESULTS
In this section, we conducted several experiments in order to evaluate the performance of the
proposed compression method. The proposed algorithm is applied over several images. The compression
efficiency and quality of the reconstructed image are compared with several other competitive image
compression techniques.
4.1. Intra Prediction
In the experiment, we used 100 images from the Berkeley segmentation database as our training set.
Nine modes of 8 x 8 intra prediction are applied and intra-prediction residual for each image is generated. A
set of 45000 blocks of size 8 x 8 are randomly selected from residual images to train a dictionary. We
selected 5000 residual blocks from each nine modes to form our learning set. Examples of intra-predicted
image and residual image are demonstrated in Figure 3. A random collection of such training blocks for
different mode is shown in Figure 4.
(a) Original Image (b) Predicted Image (c) Residual Image
Figure 3. Intra prediction of second image
(a) Mode 3 (b) Mode 6 (c) Mode 8
Figure 4. Training blocks
6. IJECE ISSN: 2088-8708
Dictionary based Image Compression via Sparse Representation (Arabinda Sahoo)
1969
4.2. K-SVD Dictionary
We used 45000 residual blocks obtained from intra-prediction as our learning set. The K-SVD
algorithm is applied to train an over-complete dictionary of size 64 x 512, where the number of rows 64
represents the number of pixels in a block, and the number of columns 512 represents the number of
dictionary elements. In the training process, 100 number of K-SVD iterations was set as the primary stopping
criterion. Sparsity constraint (T0=4) was set as another stopping criterion. Example of a trained dictionary is
demonstrated in Figure 5.
Figure 5. Example of dictionary trained on 8x8 residual blocks
4.3. Image Compression
The K-SVD Dictionary resulting from prediction residuals is used for sparse representation. In the
experiment, the residual error (δ=0.2) and the number of OMP iterations equal to 4 was set as stopping
criterion of OMP algorithm. The OMP algorithm stops when the residual error is less than or equal to 0.2,
otherwise it stops after 4 iterations. In maximum, four dictionary elements are required to encode each
residual block.
(a) (b)
(c) (d)
Figure 6. Rate-distortion curves of different methods
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2
22
24
26
28
30
32
34
36
38
40
Barbara
Bit-rate (bpp)
PSNR(dB)
JPEG
JPEG2000
Proposed Method
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2
28
30
32
34
36
38
40
42
Lena
Bit-rate (bpp)
PSNR(dB)
JPEG
JPEG2000
Proposed Method
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2
22
23
24
25
26
27
28
29
30
31
32
33
Baboon
Bit-rate (bpp)
PSNR(dB)
JPEG
JPEG2000
Proposed Method
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2
28
30
32
34
36
38
40
Peppers
Bit-rate (bpp)
PSNR(dB)
JPEG
JPEG2000
Proposed Method
7. ISSN: 2088-8708
IJECE Vol. 7, No. 4, August 2017 : 1964 – 1972
1970
In order to evaluate our proposed compression method, we compare our method with JPEG and
JPEG2000. All experiments are performed using MATLAB. The performance of proposed method is
evaluated by taking different standard test images. The quantitative evaluation of our proposed method is
accomplished using two image quality metrics: PSNR and SSIM (Structural Similarity Metric) [23]. Figure 6
shows comparison of rate-distortion curve for 4 standard test images. Table 1 shows PSNR comparison for
different images at two different bit-rates. The proposed method yields around PSNR gain of 3 dB compared
to JPEG, and PSNR gain of 0.3 dB compared to JPEG 2000. The performance in terms of average PSNR and
SSIM are shown in Table 2. The results are averaged over 9 standard test images and best results are bolded.
Subjective assessment of one image is shown in Figure 7 at bit-rate 0.2. The results show that our proposed
method outperforms JPEG and JPEG2000 in terms of all quality metrics, including PSNR and SSIM.
Table 1. PSNR (dB) comparison of the proposed method with JPG and JPEG2000 for a set of 9 test images at
two different bit-rates. Best results are bolded
PSNR(dB) at bit-rate 0.2 bpp PSNR(dB) at bit-rate 1 bpp
Images JPEG JPEG2000 Proposed method JPEG JPEG2000 Proposed method
Barbara 23.81 27.29 27.60 34.46 37.15 37.52
Sailboat 26.56 29.12 29.53 34.44 36.81 37.01
Baboon 22.48 25.24 25.45 27.98 30.65 30.81
couple 25.72 28.48 28.80 34.25 36.79 37.10
Hill 26.78 29.88 30.21 33.62 36.42 36.84
Jet plane 29.56 31.92 32.28 39.65 41.88 42.20
Lena 29.43 33.02 33.46 37.88 40.40 40.82
Lighthouse 25.82 28.39 28.82 35.50 38.42 38.83
Peppers 29.98 32.49 32.88 37.55 38.38 38.72
Average 26.68 29.54 29.89 35.03 37.43 37.76
(a) Original image (b) JPEG (c) JPEG2000 (d) Proposed method
Figure 7. Subjective assessment
Table 2. Performance comparison in terms of two image quality metrics, PSNR (dB) and SSIM at six
different bit-rates. The results are averaged over 9 test images. Best results are bolded
Quality Metric JPEG JPEG2000 Proposed method JPEG JPEG2000 Proposed
method
Bit-rate 0.2 bpp Bit-rate 0.4 bpp
PSNR 26.68 29.54 29.89 28.78 31.82 32.14
SSIM 0.748 0.771 0.782 0.762 0.844 0.861
Bit-rate 0.6 bpp Bit-rate 0.8 bpp
PSNR 31.12 33.82 34.22 33.40 36.42 36.74
SSIM 0.824 0.906 0.915 0.874 0.921 0.931
Bit-rate 1 bpp Bit-rate 1.2 bpp
PSNR 35.03 37.43 37.76 36.43 38.55 38.87
SSIM 0.909 0.941 0.949 0.918 0.951 0.954
50 100 150 200 250 300 350 400 450 500
50
100
150
200
250
300
350
400
450
500
50 100 150 200 250 300 350 400 450 500
50
100
150
200
250
300
350
400
450
500
50 100 150 200 250 300 350 400 450 500
50
100
150
200
250
300
350
400
450
500
50 100 150 200 250 300 350 400 450 500
50
100
150
200
250
300
350
400
450
500
8. IJECE ISSN: 2088-8708
Dictionary based Image Compression via Sparse Representation (Arabinda Sahoo)
1971
5. CONCLUSION
In this paper, we presented a dictionary based intra-prediction framework for image compression.
K-SVD algorithm is used in order to train a dictionary. We trained the dictionary with a variety of residual
blocks obtained from intra-prediction and then used this dictionary for sparse representation of an image.
OMP algorithm, fixed length coding, and entropy coding have employed for encoding. Different coding
results based on a set of test images are presented to compare the performance of the proposed method with
the existing methods. Experimental result shows that the proposed method outperforms JPEG and JPEG2000.
ACKNOWLEDGEMENTS
The authors are grateful to Dr. Gagan Rath for his support and guidance throughout this work.
REFERENCES
[1] Salomon, “Data Compression: The Complete Reference”, third ed., Springer, New York, 2004.
[2] W. Pennebaker, J. Mitchell, “JPEG Still Image Data Compression Standard”, Kluwer Academic Publishers, 1993.
[3] C. Christopoulos, A. Skodras, T. Ebrahimi, “The JPEG2000 Still Image Coding System: An Overview”, IEEE
Trans. Consum. Electron, 46 (2000) 1103-1127.
[4] A. Ennaciri, M. Erritali, M. Mabrouki, J. Bengourram, “Comparative Study of Wavelet Image Compression:
JPEG2000 Standart”, TELKOMNIKA Indonesian Journal of Electrical Engineering, vol. 16, no. 1, pp. 83-90, Oct.
2015.
[5] R. Chourasiya, A. Shrivastava, “A Study of Image Compression Based Transmission Algorithm using SPIHT for
Low Bit Rate Application”, Bulletin of Electrical Engineering and Informatics, vol. 2, no. 2, pp. 117-122, Jun.
2013.
[6] D. Narmadha, K. Gayathri, K. Thilagavathi, N. Sardar Basha, “An Optimal HSI Image Compression using DWT
and CP”, International Journal of Electrical and Computer Engineering, vol. 4, no. 3, pp. 411-421, Jun. 2014.
[7] M. Stirner, G. Seelmann, “Improved Redundancy Reduction for JPEG Files”, Proc. of Picture Coding Symposium
(PCS 2007), Lisbon, Portugal, Nov. 7-9, 2007.
[8] J. Yang, B. Yin, Y. Sun, N. Zhang, “A block- Matching based Intra Frame Prediction H.264/AVC”, ICME, 2006.
[9] T. Wiegand, G. J. Sullivan, G. Bjøntegaard, A. Luthra, “Overview of the H.264/AVC Video Coding Standard”,
IEEE Trans. Circuits Syst. Video Technology., vol. 13, no. 7, pp. 560-576, Jul. 2003.
[10] J. Wright, Y. Ma, J. Mairal, G. Sapiro, T. S. Huang, S. Yan, “Sparse Representation for Computer Vision and
Pattern Recognition”, Proc. IEEE, vol. 98, no. 6, pp. 1031–1044, Jun. 2010.
[11] R. Rubinstein, A. M. Bruckstein, M. Elad, “Dictionaries for Sparse Representation Modeling”, Proc. IEEE, vol.
98, no. 6, pp. 1045–1057, June 2010.
[12] J. Mairal, F. Bach, J. Ponce, and G. Sapiro, “Online dictionary learning for sparse coding,” in Proc. 26th Annu. Int.
Conf. Mach. Learning., 2009, pp. 689-696.
[13] O. Bryt, M. Elad, “Compression of Facial Images using the K-SVD Algorithm”, J. Vis. Commun. Image
Representation. vol. 19, no. 4, pp. 270-282, 2008.
[14] M. Aharon, M. Elad, A. Bruckstein, “K-SVD: An Algorithm for Designing Overcomplete Dictionaries for Sparse
Representation,” IEEE Trans. Signal Processing., vol. 54, no. 11, pp. 4311-4322, Nov. 2006.
[15] J. Zepeda, C. Guillemot, E. Kijak, “Image Compression using Sparse Representations and the Iteration-tuned and
Aligned Dictionary”, IEEE J. Sel. Topics Signal Processing, vol. 5, no. 5, pp. 1061-1073, Sep. 2011.
[16] G. Shao, Y. Wu, A. Yong, X. Liu, T. Guo, “Fingerprint Compression based on Sparse Representation”, IEEE
Trans. Image Process., vol. 23, no. 2, pp. 489-501, Feb. 2014.
[17] Jing-Ya Zhu, Zhong-Yuan Wang, Rui Zhong, Shen-Ming Qu, “Dictionary based Surveillance Image
Compression”, J. Vis. Commun. Image R. vol. 31, no. 7, pp. 225-230. Jul. 2015.
[18] G. Rath, A. Sahoo, “A Comparative Study of some Greedy Pursuit Algorithms for Sparse Approximation”,
EUSIPCO2009, pp.398-402, Aug. 2011.
[19] S. Mallat, Z. Zhang, “Matching Pursuits with Time Frequency Dictionaries”, IEEE Transactions on Signal
Processing, vol. 41, no. 12, pp. 3397-3415, Dec. 1993.
[20] Y.C.Pati, R.Rezaiifar M, P.S.Krishnaprasad, “Orthogonal Matching pursuit: Recursive Function Approximation
with Application to Wavelet Decomposition”, In Proc. 27th Conf. on Sig. Sys. and Comp., vol.1, Nov. 1993.
[21] G. Rath, C. Guillemot, “On a Simple Derivation of the Complimentary Matching Pursuit”, Signal processing,
vol. 90, no. 2, pp.7 02-706, Feb. 2010.
[22] A. Bovik, Handbook of Image and Video Processing, 2nd ed.San Francisco, CA, USA: Academic, 2005.
[23] Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli, “Image Quality Assessment: From Error Visibility to
Structural Similarity,” IEEE Trans. Image Processing., vol. 13, no. 4, pp. 600-612, Apr. 2004.
9. ISSN: 2088-8708
IJECE Vol. 7, No. 4, August 2017 : 1964 – 1972
1972
BIOGRAPHIES OF AUTHORS
Arabinda Sahoo. Currently, he is working as an Assistant Professor in the department of
Electronics & Communication Engineering at ITER, Siksha „O‟ Anusandhan University,
Bhubaneswar, Odisha, India. He has completed his B.E from Utkal University,
Bhubaneswar, Odisha, India and M.Tech from National Institute of Technology, Rourkela,
India. His research interests include signal and image processing, image compression and
sparse representations.
Pranati Das. Currently, she is working as an Associate Professor in the department of
Electrical Engineering at Indira Gandhi Institute of Technology, Sarang an Autonomous
Institute of Government of Odisha, India. She has completed her M.Tech & PhD from
Indian Institute of Technology, Kharagpur, India and Engineering graduation from
Sambalpur University, Odisha, India. She has published number of research and conference
papers in both national and international forum. Her area of research interests are Image
Processing and Signal Processing.