Self Organizing Feature Map(SOM), Topographic Product, Cascade 2 AlgorithmChenghao Jin
This document discusses several algorithms used in the SCPSNSP method for day-ahead electricity load and price forecasting, including Self Organizing Feature Map (SOM), Topographic Product, and Cascade 2. SOM is used for clustering electricity time series data into patterns. The Topographic Product measures how well the feature map represents the input data. Cascade 2, a neural network training algorithm, is used for next symbol prediction based on the clustered pattern sequences.
This document provides an overview of radiation detectors. It discusses why radiation detection is important, how radiation interacts with matter, common types of detectors like ionization chambers, proportional counters, and GM counters, and how detectors work to detect different types of radiation. Specific examples are given around using an ion chamber survey meter to detect x-rays. Key factors around detector selection, specifications, and operating principles are summarized.
This document describes a methodology for detecting brain tumors and edema from magnetic resonance images (MRI) using bounding box symmetry. The methodology involves segmenting the brain from MRI images, applying filters, and using a novel score function based on Bhattacharya coefficient to detect a rectangle representing the region of interest by comparing the left and right sides of the brain. The method was tested on 12 brain MRI images and was able to accurately detect tumor and edema regions and their sizes.
This document discusses image processing techniques in MATLAB. It begins with an introduction to MATLAB and its uses for numerical computation, data analysis, and algorithm development. It then covers image processing basics like image formats and color models. The main techniques discussed are enhancement, restoration, watermarking, cryptography, steganography, and image fusion. Examples of algorithms and real-world applications are also provided.
Nd:YAG lasers emit light at a wavelength of 1.064 μm. They use neodymium-doped yttrium aluminum garnet (Nd:YAG) as the active medium and are optically pumped by a xenon flash lamp. Nd:YAG lasers can operate in pulsed mode and produce up to 20 kW of power output. They use an optical resonator formed between partially and fully reflective mirrors to produce stimulated emission along the length of the Nd:YAG rod. Lasers have applications in fields like medicine, communications, industry, science and technology, and the military.
The document summarizes key concepts in image formation, including how light interacts with objects and lenses to form images, and how different imaging systems like the human eye and digital cameras work. It discusses factors that affect image quality such as point spread functions and noise. Methods for analyzing the effects of noise propagation and algorithms on image quality are presented, such as error propagation techniques and Monte Carlo simulations.
The document discusses the Discrete Fourier Transform (DFT). It begins by explaining the limitations of the Discrete Time Fourier Transform (DTFT) and Discrete Fourier Series (DFS) from a numerical computation perspective. It then introduces the DFT as a numerically computable transform obtained by sampling the DTFT in the frequency domain. The DFT represents a periodic discrete-time signal using a sum of complex exponentials. It defines the DFT and inverse DFT equations. The document also discusses properties of the DFT such as linearity and time/frequency shifting. Finally, it notes that the Fast Fourier Transform (FFT) implements the DFT more efficiently by constraining the number of points to powers of two.
This document discusses digital image processing concepts including:
- Image acquisition and representation, including sampling and quantization of images. CCD arrays are commonly used in digital cameras to capture images as arrays of pixels.
- A simple image formation model where the intensity of a pixel is a function of illumination and reflectance at that point. Typical ranges of illumination and reflectance are provided.
- Image interpolation techniques like nearest neighbor, bilinear, and bicubic interpolation which are used to increase or decrease the number of pixels in a digital image. Examples of applying these techniques are shown.
- Basic relationships between pixels including adjacency, paths, regions, boundaries, and distance measures like Euclidean, city block, and
This document discusses using MATLAB to detect breast cancer through analysis of mammogram and thermal images. It introduces breast cancer and explains that early detection is key to successful treatment. Currently, mammography and thermography are used for detection, but mammography has weaknesses like pain and radiation. The purpose of this project is to design a system to detect signs in mammogram and thermal images using image processing techniques in MATLAB. Mammogram images will be analyzed using morphology before feature extraction and classification. Thermal images will have features extracted from the heat distribution to identify possible cancer areas.
This document describes the syllabus for a biomedical optics course. It covers 13 topics including Monte Carlo modeling of photon transport, the radiative transfer equation, diffuse optical tomography, and ultrasound-modulated optical tomography. The first chapter introduces motivation for optical imaging in biology, general behavior of light in tissue, and basic light-matter interactions. It discusses optical properties of tissue, Beer's law, absorption and scattering origins and coefficients, and exercises calculating light transmission through absorbing and scattering media. Oxygen saturation and concentration are also introduced.
This document discusses radiation monitoring instruments used for external exposure monitoring. It describes how area survey meters are used to measure radiation levels in work areas and around radiotherapy equipment. These include ionization chambers, proportional counters, and Geiger-Muller counters, which use gas detectors to measure different types of radiation. Personal dosimeters are also discussed, which are worn by individuals to record equivalent radiation doses. All monitoring instruments must be calibrated to appropriate radiation protection quantities.
This document discusses self-organizing maps (SOM), an unsupervised machine learning technique that projects high-dimensional data into a low-dimensional space. SOM creates a map that clusters similar data items together and separates dissimilar items. It is useful for data mining, data analysis, and pattern recognition. The document provides examples of using SOM to cluster metallic elements based on their physical properties and cluster different soil types based on their spectral properties with increasing noise.
The document discusses various activation functions used in deep learning neural networks including sigmoid, tanh, ReLU, LeakyReLU, ELU, softmax, swish, maxout, and softplus. For each activation function, the document provides details on how the function works and lists pros and cons. Overall, the document provides an overview of common activation functions and considerations for choosing an activation function for different types of deep learning problems.
This document discusses region-based image segmentation techniques. Region-based segmentation groups pixels into regions based on common properties. Region growing is described as starting with seed points and grouping neighboring pixels with similar properties into larger regions. The advantages are it can correctly separate regions with the same defined properties and provide good segmentation in images with clear edges. The disadvantages include being computationally expensive and sensitive to noise. Region splitting and merging techniques are also discussed as alternatives to region growing.
The document discusses digital image processing. It begins by defining an image and describing how images are represented digitally. It then outlines the main steps in digital image processing, including acquisition, enhancement, restoration, segmentation, representation, and recognition. It also discusses the key components of an image processing system, including hardware, software, storage, displays, and networking. Finally, it provides examples of application areas for digital image processing such as medical imaging, satellite imaging, and industrial inspection.
The document summarizes key aspects of a cyclotron, which is a device that accelerates charged particles outwards in a spiral path using crossed electric and magnetic fields. It was invented in 1929 and the first operational cyclotron was built in 1932 by Ernest Lawrence. Cyclotrons work by subjecting particles to an oscillating electric field while they travel in a circle due to a static magnetic field. Modifications allow relativistic speeds. Cyclotrons are used in nuclear physics experiments and for producing isotopes for PET imaging and particle cancer therapy. Limitations include inability to accelerate neutral particles or electrons.
In multirate digital signal processing, the sampling rate of a signal is changed in order to increase the efficiency of various signal processing operations. This technique is used in audio CD, where the sampling frequency of 44.1 kHz is increased fourfold to 176.4 kHz before D/A conversion
Object Detection on Dental X-ray Images using R-CNNMinhazul Arefin
The document describes an object detection model using region-based convolutional neural networks to detect dental caries and root canals in dental x-ray images. The model was trained on a dataset of dental x-rays labeled with caries and root canal objects. Feature extraction was performed using multiple convolutional layers and concatenation. Anchors, multi-scale training, and hard negative mining were used to improve performance. The model achieved 83.45% accuracy for object detection, outperforming other methods. Future work could include working with larger datasets and real-time detection from video.
This document discusses the cyclotron, a type of particle accelerator. It begins with an introduction and overview of key topics like principles, construction, diagrams, workings, calculations, applications, and limitations. Some key points made are:
- A cyclotron accelerates charged particles like protons and deuterons using electric and magnetic fields, generating energies from 1 MeV to over 100 MeV.
- It works on the principle that a charged particle moving perpendicular to a magnetic field experiences a force causing it to travel in a circular path, with increasing radius and velocity over time due to an oscillating electric field.
- Important applications of cyclotrons include production of beams for nuclear physics experiments and cancer particle therapy.
Digital Signal Processing Lab Manual ECE studentsUR11EC098
This document describes a MATLAB program to perform operations on discrete-time signals. It discusses amplitude manipulation operations like amplification, attenuation, and amplitude reversal. Time manipulation operations covered include time shifting and time reversal. It also describes adding and multiplying two discrete signals. The program takes user input, performs the selected operations, and plots the output waveforms to verify results.
The document describes the proportional counter, which is a gaseous state particle detector used to detect nuclear particles and radiation. It consists of a cylindrical metal tube filled with argon and methane gas and a thin metal wire running down the center as an anode. When radiation enters the tube, it ionizes the gas, producing electron-ion pairs. An applied voltage between the wire and tube causes gas amplification through avalanching, resulting in a pulse signal. The proportional counter can be used for particle counting and energy determination, and has advantages like low-energy detection but requires stable applied voltages.
1) The document discusses finite word length effects in digital filters. It covers fixed point and floating point number representations, different number systems including binary, decimal, octal and hexadecimal.
2) It describes various number representation techniques for digital systems including fixed point representation, floating point representation, and block floating point representation. Fixed point representation uses a fixed binary point position while floating point representation allows the binary point to vary.
3) It also discusses signed number representations including sign-magnitude, one's complement, and two's complement forms. Arithmetic operations like addition, subtraction and multiplication are covered for fixed point numbers along with issues like overflow.
This document discusses different types of solid state radiation detectors, including scintillation detectors, thermoluminescent dosimeters (TLD), and semiconductor detectors. Scintillation detectors detect radiation via light emission in inorganic crystal materials like NaI or organic crystals like anthracene. TLDs "capture" radiation dose information and release light when heated, allowing dose measurement. Common TLD materials are LiF:Mg,Ti and Li2B4O7:Mn. Semiconductor detectors like silicon and germanium act as solid state ionization chambers and are used for high resolution energy measurement of alpha and beta particles.
This document discusses scintillation detectors and their properties. Scintillation detectors work by emitting light when exposed to radiation, and this light output can be used to measure incident radiation. The key properties of scintillation detectors are that they must have high scintillation efficiency, light yield proportional to deposited energy, short decay time, transparency to emitted wavelengths, and be able to be made in large sizes and desired shapes. Common inorganic scintillators discussed are NaI(Tl), which is widely used due to its availability and high detection efficiency, and BGO, which has high intrinsic efficiency for high gamma energies.
The document discusses the interaction of radiation with matter. It explains that radiation can be electromagnetic or particulate. When electromagnetic radiation like x-rays or gamma rays pass through matter, they can undergo attenuation, absorption, scattering, or transmission. The major interactions that cause attenuation are coherent scattering, the photoelectric effect, the Compton effect, pair production, and photonuclear interactions. It describes each of these interactions in detail and how they transfer energy from the radiation to the absorbing material.
This document discusses various types of radiation detection devices, including film badges, ionization chambers, Geiger-Muller counters, proportional counters, scintillation counters, photographic plates, electroscopes, bubble chambers, solid-state detectors, cloud chambers, and spark counters. Each detection method works by using different processes like ionization, fluorescence, or track visualization to detect and sometimes quantify radiation levels or particle energy. Regular monitoring of radiation is important for safety when working with radioactive materials.
Bioinformatics Project Training for 2,4,6 monthbiinoida
The document discusses projects and research opportunities available at the Bioinformatics Institute of India (BII). Some key points:
- BII focuses on major bioinformatics areas like genome analysis, protein structure prediction, microarray analysis, and drug discovery.
- Students can complete short or long-term projects at BII on topics like sequence analysis, phylogenetics, and software development.
- Projects help students gain skills in bioinformatics tools like MATLAB, Bioperl, and develop their careers in fields like academia, industry, and research.
- BII has hosted projects for students from various universities on diverse topics ranging from protein structure prediction to disease research.
The document provides an overview of self-organizing maps (SOM). It defines SOM as an unsupervised learning technique that reduces the dimensions of data through the use of self-organizing neural networks. SOM is based on competitive learning where the closest neural network unit to the input vector (the best matching unit or BMU) is identified and adjusted along with neighboring units. The algorithm involves initializing weight vectors, presenting input vectors, identifying the BMU, and updating weights of the BMU and neighboring units. SOM can be used for applications like dimensionality reduction, clustering, and visualization.
This document discusses using MATLAB to detect breast cancer through analysis of mammogram and thermal images. It introduces breast cancer and explains that early detection is key to successful treatment. Currently, mammography and thermography are used for detection, but mammography has weaknesses like pain and radiation. The purpose of this project is to design a system to detect signs in mammogram and thermal images using image processing techniques in MATLAB. Mammogram images will be analyzed using morphology before feature extraction and classification. Thermal images will have features extracted from the heat distribution to identify possible cancer areas.
This document describes the syllabus for a biomedical optics course. It covers 13 topics including Monte Carlo modeling of photon transport, the radiative transfer equation, diffuse optical tomography, and ultrasound-modulated optical tomography. The first chapter introduces motivation for optical imaging in biology, general behavior of light in tissue, and basic light-matter interactions. It discusses optical properties of tissue, Beer's law, absorption and scattering origins and coefficients, and exercises calculating light transmission through absorbing and scattering media. Oxygen saturation and concentration are also introduced.
This document discusses radiation monitoring instruments used for external exposure monitoring. It describes how area survey meters are used to measure radiation levels in work areas and around radiotherapy equipment. These include ionization chambers, proportional counters, and Geiger-Muller counters, which use gas detectors to measure different types of radiation. Personal dosimeters are also discussed, which are worn by individuals to record equivalent radiation doses. All monitoring instruments must be calibrated to appropriate radiation protection quantities.
This document discusses self-organizing maps (SOM), an unsupervised machine learning technique that projects high-dimensional data into a low-dimensional space. SOM creates a map that clusters similar data items together and separates dissimilar items. It is useful for data mining, data analysis, and pattern recognition. The document provides examples of using SOM to cluster metallic elements based on their physical properties and cluster different soil types based on their spectral properties with increasing noise.
The document discusses various activation functions used in deep learning neural networks including sigmoid, tanh, ReLU, LeakyReLU, ELU, softmax, swish, maxout, and softplus. For each activation function, the document provides details on how the function works and lists pros and cons. Overall, the document provides an overview of common activation functions and considerations for choosing an activation function for different types of deep learning problems.
This document discusses region-based image segmentation techniques. Region-based segmentation groups pixels into regions based on common properties. Region growing is described as starting with seed points and grouping neighboring pixels with similar properties into larger regions. The advantages are it can correctly separate regions with the same defined properties and provide good segmentation in images with clear edges. The disadvantages include being computationally expensive and sensitive to noise. Region splitting and merging techniques are also discussed as alternatives to region growing.
The document discusses digital image processing. It begins by defining an image and describing how images are represented digitally. It then outlines the main steps in digital image processing, including acquisition, enhancement, restoration, segmentation, representation, and recognition. It also discusses the key components of an image processing system, including hardware, software, storage, displays, and networking. Finally, it provides examples of application areas for digital image processing such as medical imaging, satellite imaging, and industrial inspection.
The document summarizes key aspects of a cyclotron, which is a device that accelerates charged particles outwards in a spiral path using crossed electric and magnetic fields. It was invented in 1929 and the first operational cyclotron was built in 1932 by Ernest Lawrence. Cyclotrons work by subjecting particles to an oscillating electric field while they travel in a circle due to a static magnetic field. Modifications allow relativistic speeds. Cyclotrons are used in nuclear physics experiments and for producing isotopes for PET imaging and particle cancer therapy. Limitations include inability to accelerate neutral particles or electrons.
In multirate digital signal processing, the sampling rate of a signal is changed in order to increase the efficiency of various signal processing operations. This technique is used in audio CD, where the sampling frequency of 44.1 kHz is increased fourfold to 176.4 kHz before D/A conversion
Object Detection on Dental X-ray Images using R-CNNMinhazul Arefin
The document describes an object detection model using region-based convolutional neural networks to detect dental caries and root canals in dental x-ray images. The model was trained on a dataset of dental x-rays labeled with caries and root canal objects. Feature extraction was performed using multiple convolutional layers and concatenation. Anchors, multi-scale training, and hard negative mining were used to improve performance. The model achieved 83.45% accuracy for object detection, outperforming other methods. Future work could include working with larger datasets and real-time detection from video.
This document discusses the cyclotron, a type of particle accelerator. It begins with an introduction and overview of key topics like principles, construction, diagrams, workings, calculations, applications, and limitations. Some key points made are:
- A cyclotron accelerates charged particles like protons and deuterons using electric and magnetic fields, generating energies from 1 MeV to over 100 MeV.
- It works on the principle that a charged particle moving perpendicular to a magnetic field experiences a force causing it to travel in a circular path, with increasing radius and velocity over time due to an oscillating electric field.
- Important applications of cyclotrons include production of beams for nuclear physics experiments and cancer particle therapy.
Digital Signal Processing Lab Manual ECE studentsUR11EC098
This document describes a MATLAB program to perform operations on discrete-time signals. It discusses amplitude manipulation operations like amplification, attenuation, and amplitude reversal. Time manipulation operations covered include time shifting and time reversal. It also describes adding and multiplying two discrete signals. The program takes user input, performs the selected operations, and plots the output waveforms to verify results.
The document describes the proportional counter, which is a gaseous state particle detector used to detect nuclear particles and radiation. It consists of a cylindrical metal tube filled with argon and methane gas and a thin metal wire running down the center as an anode. When radiation enters the tube, it ionizes the gas, producing electron-ion pairs. An applied voltage between the wire and tube causes gas amplification through avalanching, resulting in a pulse signal. The proportional counter can be used for particle counting and energy determination, and has advantages like low-energy detection but requires stable applied voltages.
1) The document discusses finite word length effects in digital filters. It covers fixed point and floating point number representations, different number systems including binary, decimal, octal and hexadecimal.
2) It describes various number representation techniques for digital systems including fixed point representation, floating point representation, and block floating point representation. Fixed point representation uses a fixed binary point position while floating point representation allows the binary point to vary.
3) It also discusses signed number representations including sign-magnitude, one's complement, and two's complement forms. Arithmetic operations like addition, subtraction and multiplication are covered for fixed point numbers along with issues like overflow.
This document discusses different types of solid state radiation detectors, including scintillation detectors, thermoluminescent dosimeters (TLD), and semiconductor detectors. Scintillation detectors detect radiation via light emission in inorganic crystal materials like NaI or organic crystals like anthracene. TLDs "capture" radiation dose information and release light when heated, allowing dose measurement. Common TLD materials are LiF:Mg,Ti and Li2B4O7:Mn. Semiconductor detectors like silicon and germanium act as solid state ionization chambers and are used for high resolution energy measurement of alpha and beta particles.
This document discusses scintillation detectors and their properties. Scintillation detectors work by emitting light when exposed to radiation, and this light output can be used to measure incident radiation. The key properties of scintillation detectors are that they must have high scintillation efficiency, light yield proportional to deposited energy, short decay time, transparency to emitted wavelengths, and be able to be made in large sizes and desired shapes. Common inorganic scintillators discussed are NaI(Tl), which is widely used due to its availability and high detection efficiency, and BGO, which has high intrinsic efficiency for high gamma energies.
The document discusses the interaction of radiation with matter. It explains that radiation can be electromagnetic or particulate. When electromagnetic radiation like x-rays or gamma rays pass through matter, they can undergo attenuation, absorption, scattering, or transmission. The major interactions that cause attenuation are coherent scattering, the photoelectric effect, the Compton effect, pair production, and photonuclear interactions. It describes each of these interactions in detail and how they transfer energy from the radiation to the absorbing material.
This document discusses various types of radiation detection devices, including film badges, ionization chambers, Geiger-Muller counters, proportional counters, scintillation counters, photographic plates, electroscopes, bubble chambers, solid-state detectors, cloud chambers, and spark counters. Each detection method works by using different processes like ionization, fluorescence, or track visualization to detect and sometimes quantify radiation levels or particle energy. Regular monitoring of radiation is important for safety when working with radioactive materials.
Bioinformatics Project Training for 2,4,6 monthbiinoida
The document discusses projects and research opportunities available at the Bioinformatics Institute of India (BII). Some key points:
- BII focuses on major bioinformatics areas like genome analysis, protein structure prediction, microarray analysis, and drug discovery.
- Students can complete short or long-term projects at BII on topics like sequence analysis, phylogenetics, and software development.
- Projects help students gain skills in bioinformatics tools like MATLAB, Bioperl, and develop their careers in fields like academia, industry, and research.
- BII has hosted projects for students from various universities on diverse topics ranging from protein structure prediction to disease research.
The document provides an overview of self-organizing maps (SOM). It defines SOM as an unsupervised learning technique that reduces the dimensions of data through the use of self-organizing neural networks. SOM is based on competitive learning where the closest neural network unit to the input vector (the best matching unit or BMU) is identified and adjusted along with neighboring units. The algorithm involves initializing weight vectors, presenting input vectors, identifying the BMU, and updating weights of the BMU and neighboring units. SOM can be used for applications like dimensionality reduction, clustering, and visualization.
The document discusses unsupervised learning and the self-organizing map (SOM) algorithm. The SOM is inspired by biological neural maps and organizes high-dimensional input data onto a low-dimensional grid while preserving topological properties. The algorithm works by finding the best matching unit on the grid for each input and adjusting its weights and those of nearby units. SOMs can be used to cluster multidimensional data and visualize relationships that may otherwise be difficult to detect. They are proposed as a way to cluster agricultural sites based on multiple environmental characteristics to determine suitable crops and varieties for different locations.
IEEE- Intrusion Detection Model using Self Organizing MapTushar Shinde
This document proposes an intrusion detection model using self-organizing maps (SOM) to detect malicious attackers on websites. It discusses how denial of service (DoS) attacks aim to harm systems by flooding servers with traffic. The proposed model uses an unsupervised machine learning technique called SOM to analyze website authentication logs and achieve better security by detecting malicious attackers. The SOM algorithm is chosen to naturally cluster data and produce better results than other clustering algorithms. Pseudocode is provided to demonstrate how the SOM algorithm is implemented on normalized website log data to identify different types of visitors and detect intrusions.
The document proposes a resource allocation strategy using optimal power control to mitigate interference in heterogeneous networks. It presents a system model where macrocells and femtocells are deployed. Users are divided into cell center and cell edge groups. A soft fractional frequency reuse scheme is used to improve cell edge performance. Simulations show the proposed approach improves coverage probability and SINR compared to other schemes, especially for cell edge users.
Timo Honkela: Kohonen's Self-Organizing Maps for Intelligent Systems Developm...Timo Honkela
An invited talk given in the FODO'98, Foundations of Data Organization conference. The conference took place in Kobe, Japan, November 12-13, 1998. Main themes of the talk included Self-Organizing Maps (SOMs), Fuzzy Sets, context analysis, and systems of SOMs.
Timo Honkela: Turning quantity into quality and making concepts visible using...Timo Honkela
Professor Timo Honkela gave an invited talk in the Göran Mickwitz seminar that took place in Helsinki, 9th of February 2017. The event was organized in the honor of Doc. Jessica Parland-von Essen.
The document discusses a clinical decision support system (CDSS) being developed as part of the Synergy-COPD project, which will use a Java-based framework and Drools rules engine to represent clinical knowledge and make inferences from patient data represented in a HL7 virtual medical record format, with the goal of aiding diagnosis and management of COPD patients.
This document discusses gene prediction and some of the computational challenges involved. It covers topics like genes and proteins, gene prediction problems, computational approaches like using open reading frames and codon usage. It also discusses central dogma, exons, introns, splicing signals, genetic code and different gene prediction algorithms.
1. The document discusses a talk on using ontologies like Gene Ontology (GO) to improve the data mining step of knowledge discovery from databases.
2. It introduces IntelliGO, a semantic similarity measure for GO annotations that can be used for gene clustering and abstraction of biomedical data to aid data mining.
3. The talk will cover the development of IntelliGO, its application to gene clustering, and its use in clustering secondary effects from another ontology to reduce data dimensions for mining.
Timo Honkela: Self-Organizing Map as a Means for Gaining PerspectivesTimo Honkela
Presentation on 23rd of May, 2014, in Metalithicum # 5, Computation as literacy: Self Organizing Maps, organized by ETH CAAD in Einsiedeln, Switzerland.
Program:
Thursday 22nd of May 2014
13:30-14:30 INTRODUCTION – CODING AND ARCHITECTURE
Prof. Dr. Ludger Hovestadt
Chair for Computer Aided Architectural Design, CAAD, ITA, ETH Zurich
14:30-16:00 Discussion
17:30-18:30 WARM UP TWO – PROFILING KEY CONCEPTS IN CONTINUOUS GEOMETRY
Prof. Sha Xin Wei
Director School of Arts, Media and Engineering, Herberger Institute
for Design and the Arts, Arizona State University, Founding Director
Topological Media Lab, Concordia University, Montreal.
18:30-19:00 Discussion
Friday 23rd of May 2014
08:00-09:00 WARM UP – PROFILING KEY CONCEPTS IN CATEGORY
THEORY
Prof. Michael Epperson
Center for Philosophy and the natural Science, College of Natural
Sciences and Mathematics, California State University, Sacramento, USA
10:00-10.30 Discussion
10:30-11:30 SELF-ORGANIZING MAP AS A MEANS FOR GAINING PERSPECTIVES
Prof. Dr. Timo Honkela
Department of Modern Language, University of Helsinki and
National Library of Finland
11:30-12:30 Discussion
13:00-14:00 Prof. Barbara Hammer
CITEC centre of excellence, Bielefeld University, Bielefeld, Germany
14:00-15:00 Discussion
15:30-16:30 THE PRACTICAL PROBLEM OF CALIBRATING TOPOLOGICAL
DYNAMICS AGAINST SOCIO-CULTURAL & HISTORICAL PROCESSES
Prof. Dr. Sha Xin Wei
Director School of Arts, Media and Engineering, Herberger Institute
for Design and the Arts, Arizona State University, Founding Director
Topological Media Lab, Concordia University, Montreal
16:30-17:30 Discussion
18:00-19:00 Dr. Elias Zafiris
Department of Mathematics at the University of Athens
19:00-20:00 Discussion
Saturday 24th of May 2014
9:00-10:00 Dr. André Skupin
Department of Geography San Diego State University,
http://geography.sdsu.edu/People/Pages/skupin/
10:00-11:00 Discussion
11:30-12:30 Vahid Moosavi
PhD Candidate at the Chair for Computer Aided Architectural Design,
CAAD, ITA, ETH Zurich, www.caad.arch.ethz.ch, Researcher at Future
Cities Laboratory, Singapore-ETH Centre
12:30-13:30 Discussion
14:30-15:30 THE ONTOLOGY AND EPISTEMOLOGY OF INTERNAL RELATIONS:
BRIDGING THE PHYSICAL AND CONCEPTUAL IN QUANTUM
MECHANICS AND QUANTUM INFORMATION
Prof. Dr. Michael Epperson
Center for Philosophy and the natural Science, College of Natural
Sciences and Mathematics, California State University, Sacramento, USA
15:30-16:30 Discussion
17:00-18:00 Dr. phil. Vera Bühlmann
laboratory for applied virtuality, Chair for Computer Aided
Architectural Design, CAAD, ITA, ETH Zurich
18:00-19:00 Discussion
"k-means-clustering" presentation @ Papers We Love BucharestAdrian Florea
This document provides an overview of k-means clustering, an unsupervised machine learning technique for grouping unlabeled data points into clusters based on their similarities. It describes how k-means works by initializing cluster centroids and then iteratively reassigning data points to the closest centroid and recalculating centroids until clusters stabilize. The document also discusses pros and cons of k-means clustering as well as tips, tools, and references for further reading.
Similarity-based gene prediction approaches compare unknown genes in a genome to known genes in closely-related species to determine gene structure and function. A key challenge is that genes are split into exons and introns. Spliced alignment algorithms aim to find the best chain of exons in the genome that align to a target gene sequence, taking into account local alignments between potential exons. Popular gene prediction tools that use spliced alignment include GENSCAN, GenomeScan, TwinScan, Glimmer, and GenMark. They apply probabilistic models and similarity information to predict exon-intron structure.
This document discusses gene and DNA sequence patenting as well as patenting transgenic organisms. It defines what a patent is and explains that genes, DNA sequences, and transgenic organisms can be patented if they are novel, non-obvious, and industrially applicable. However, gene patenting is controversial from an ethics perspective as some believe genes should not be treated as property. Examples of patented genes, transgenic mice and fish are provided.
Gene expression in eukaryotes is controlled at multiple levels, including chromatin structure, transcription, RNA processing, and translation. Chromatin structure determines if genes are transcriptionally active or inactive. Transcription is regulated by the interaction of promoters, transcription factors, and enhancers. RNA processing controls splicing and transport of mRNA. Finally, translation and post-translational modifications further regulate gene expression. Overall, eukaryotic gene expression is tightly controlled through complex mechanisms at the chromatin, transcription, RNA, translation, and protein levels.
Neural networks Self Organizing Map by Engr. Edgar Carrillo IIEdgar Carrillo
This presentation talks about neural networks and self organizing maps. In this presentation,Engr. Edgar Caburatan Carrillo II also discusses its applications.
This document provides an overview of self-organizing maps (SOM) as an unsupervised learning technique. It discusses the principles of self-organization including self-amplification, competition, and cooperation. The Willshaw-von der Malsburg model and Kohonen feature maps are presented as two approaches to building topographic maps through self-organization. The Kohonen SOM learning algorithm is described as involving competition between neurons to determine a winning neuron, cooperation between neighboring neurons, and adaptive changes to synaptic weights based on Hebbian learning principles.
Secondary Structure Prediction of proteins Vijay Hemmadi
Secondary structure prediction has been around for almost a quarter of a century. The early methods suffered from a lack of data. Predictions were performed on single sequences rather than families of homologous sequences, and there were relatively few known 3D structures from which to derive parameters. Probably the most famous early methods are those of Chou & Fasman, Garnier, Osguthorbe & Robson (GOR) and Lim. Although the authors originally claimed quite high accuracies (70-80 %), under careful examination, the methods were shown to be only between 56 and 60% accurate (see Kabsch & Sander, 1984 given below). An early problem in secondary structure prediction had been the inclusion of structures used to derive parameters in the set of structures used to assess the accuracy of the method.
Some good references on the subject:
This document discusses correlation analysis in time and space using random fields and field meta-models. It provides examples of:
1) Parameterizing a dynamic process using random fields to model variations over time with only a few parameters.
2) Using field meta-models to perform sensitivity analysis on signals and identify which inputs most influence variation at different points in time.
3) Applying these methods to spatial variations, such as modeling geometric imperfections based on laser scans to generate random designs for robustness analysis.
Coates p: the use of genetic programing in exploring 3 d design worldsArchiLab 7
This paper discusses two projects using genetic programming to generate 3D forms. The first project uses an interactive genetic programming approach to evolve a shape grammar by having a user select shapes for reproduction. The second project uses a genetic algorithm to evolve Lindenmayer system rule sets to achieve a target 3D configuration, with the Hausdorff dimension driving selection. The paper provides details on how genetic programming was implemented for both projects, including the use of functions, terminals, initial populations, and fitness evaluations.
A Comparison of Traditional Simulation and MSAL (6-3-2015)Bob Garrett
This document compares traditional simulation approaches to the Model-Simulation-Analysis-Looping (MSAL) approach. It provides background information on system modeling and simulation basics, including conceptual models, simulation programs, sensitivity analysis, Monte Carlo methods, and simulation optimization. It then discusses risk and uncertainty, modeling systems of systems, and the current state of modeling and simulation in systems engineering. Finally, it introduces the MSAL approach, which uses graphs, analytics, and repeated simulation loops to address the increased complexity and uncertainty in systems of systems compared to traditional approaches. The MSAL approach aims to provide benefits like improved handling of uncertainty and complexity.
Rauber, a. 1999: label_som_on the labeling of self-organising mapsArchiLab 7
This document introduces the LabelSOM approach for automatically labeling the units of a self-organizing map (SOM) based on the characteristics learned during training. The LabelSOM method labels each unit with the input features that best characterize the data points mapped to that unit. This provides informative labels that describe the clusters in an interpretable way. The document demonstrates LabelSOM on two datasets: 1) an animal attributes dataset, and 2) a dataset of scientific publication abstracts, generating a labeled map of research topics.
Face Recognition System using Self Organizing Feature Map and Appearance Base...ijtsrd
Face Recognition has develop one of the most effective presentations of image analysis. This area of research is important not only for the applications in human computer interaction, biometric and security but also in other pattern classification problem. To improve face recognition in this system, two methods are used PCA Principal component analysis and SOM Self organizing feature Map .PCA is a subspace projection method is used compress the input face image. SOM method is used to classify DCT based feature vectors into groups to identify if the subject in the input image is "present" or "not present" in the image database. The aim of this system is that input image has to compare with stored images in the database using PCA and SOM method. An image database of 100 face images is evaluated containing 10 subjects and each subject having 10 images with different facial expression. This system is evaluated by measuring the accuracy of recognition rate. This system has been implemented by MATLAB programming. Thaung Yin | Khin Moh Moh Than | Win Tun "Face Recognition System using Self-Organizing Feature Map and Appearance-Based Approach" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd26691.pdfPaper URL: https://www.ijtsrd.com/computer-science/cognitive-science/26691/face-recognition-system-using-self-organizing-feature-map-and-appearance-based-approach/thaung-yin
Instruction level parallelism using ppm branch predictionIAEME Publication
This document summarizes an approach to instruction level parallelism using prediction by partial matching (PPM) branch prediction. It proposes a hybrid PPM-based branch predictor that uses both local and global branch histories. The two predictors are combined using a neural network. Key aspects of the implementation include:
1. Using local and global history PPM predictors and combining their predictions with a neural network.
2. Enhancements to the basic PPM approach like program counter tagging, efficient history encoding using run-length encoding, tracking pattern bias, and dynamic pattern length selection.
3. Details of the global history PPM predictor including the use of tables and linked lists to store patterns of different lengths and handle collisions
This document discusses machine learning algorithms and their suitability for parallelization using MapReduce. It begins by introducing machine learning and different types of algorithms, including supervised, unsupervised, and reinforcement learning. It then discusses how several common machine learning algorithms can be expressed using the MapReduce framework. Single-pass algorithms like language modeling and naive Bayes classification are well-suited since they involve extracting statistics from each data point. Iterative algorithms can also be parallelized by chaining multiple MapReduce jobs, though the map tasks may need access to global parameters. Overall, the document analyzes how different machine learning algorithms map to the data processing patterns of MapReduce.
This paper studies a new, quantitative approach using fractal geometry to analyse basic tenets
of good programming style. Experiments on C source of the GNU/Linux Core Utilities, a
collection of 114 programs or approximately 70,000 lines of code, show systematic changes in
style are correlated with statistically significant changes in fractal dimension (P≤0.0009). The
data further show positive but weak correlation between lines of code and fractal dimension
(r=0.0878). These results suggest the fractal dimension is a reliable metric of changes that
affect good style, the knowledge of which may be useful for maintaining a code base.
FRACTAL ANALYSIS OF GOOD PROGRAMMING STYLEcscpconf
This paper studies a new, quantitative approach using fractal geometry to analyze basic tenets of good programming style. Experiments on C source of the GNU/Linux Core Utilities, a
collection of 114 programs or approximately 70,000 lines of code, show systematic changes in style are correlated with statistically significant changes in fractal dimension (P≤0.0009). The data further show positive but weak correlation between lines of code and fractal dimension (r=0.0878). These results suggest the fractal dimension is a reliable metric of changes that
affect good style, the knowledge of which may be useful for maintaining a code base.
This document provides an introduction and overview of SmartPLS software for structural equation modeling. It discusses key concepts in SEM like latent and manifest variables, reflective and formative measurement models, and the structural and measurement models. It then demonstrates how to use SmartPLS by opening a project file, evaluating a sample research model and hypotheses, and interpreting the output metrics to assess model fit and quality.
Probabilistic Programming for Dynamic Data Assimilation on an Agent-Based ModelNick Malleson
Usually in computer programming, variables are usually assigned to specific values (e.g. a virtual person in a computer simulation might have an 'age' variable which stores a number). Probabilistic programming, on the other hand, allows you to represent _random variables_. These are variables whose values we do not know precisely, so can be represented by a probability distribution rather than a single value.
This is potentially a very elegant way of capturing uncertainty in our models. However, using the probabilistic programming approach for agent-based modelling raises a number of questions. In the following slides, LIDA data science intern Luke Archer introduces his recent work that explores the use of probabilistic programming libraries for building agent-based models.
GENETIC ALGORITHM FOR FUNCTION APPROXIMATION: AN EXPERIMENTAL INVESTIGATIONijaia
Function Approximation is a popular engineering problems used in system identification or Equation
optimization. Due to the complex search space it requires, AI techniques has been used extensively to spot
the best curves that match the real behavior of the system. Genetic algorithm is known for their fast
convergence and their ability to find an optimal structure of the solution. We propose using a genetic
algorithm as a function approximator. Our attempt will focus on using the polynomial form of the
approximation. After implementing the algorithm, we are going to report our results and compare it with
the real function output.
A Framework for Visualizing Association Mining Resultsertekg
Download Link > https://ertekprojects.com/gurdal-ertek-publications/blog/a-framework-for-visualizing-association-mining-results/
Association mining is one of the most used data mining techniques due to interpretable and actionable results. In this study we pro-pose a framework to visualize the association mining results, speci¯cally frequent itemsets and association rules, as graphs. We demonstrate the applicability and usefulness of our approach through a Market Basket Analysis (MBA) case study where we visually explore the data mining results for a supermarket data set. In this case study we derive several
interesting insights regarding the relationships among the items and sug-gest how they can be used as basis for decision making in retailing.
A novel hybrid deep learning model for price prediction IJECEIAES
Price prediction has become a major task due to the explosive increase in the number of investors. The price prediction task has various types such as shares, stocks, foreign exchange instruments, and cryptocurrency. The literature includes several models for price prediction that can be classified based on the utilized methods into three main classes, namely, deep learning, machine learning, and statistical. In this context, we proposed several models’ architectures for price prediction. Among them, we proposed a hybrid one that incorporates long short-term memory (LSTM) and Convolution neural network (CNN) architectures, we called it CNN-LSTM. The proposed CNNLSTM model makes use of the characteristics of the convolution layers for extracting useful features embedded in the time series data and the ability of LSTM architecture to learn long-term dependencies. The proposed architectures are thoroughly evaluated and compared against state-of-the-art methods on three different types of financial product datasets for stocks, foreign exchange instruments, and cryptocurrency. The obtained results show that the proposed CNN-LSTM has the best performance on average for the utilized evaluation metrics. Moreover, the proposed deep learning models were dominant in comparison to the state-of-the-art methods, machine learning models, and statistical models.
The document discusses Kohonen self-organizing maps (SOMs), a type of neural network developed by Tuevo Kohonen in 1982 that are able to map high-dimensional input data to lower dimensional spaces like two dimensions through unsupervised competitive learning. SOMs find application in areas like text clustering, medical imaging analysis, and classifying world poverty by organizing large, complex datasets in an interpretable manner, though determining initial input weights and requiring large training datasets are disadvantages.
This document summarizes principal component analysis (PCA) and its application to face recognition. PCA is a technique used to reduce the dimensionality of large datasets while retaining the variations present in the dataset. It works by transforming the dataset into a new coordinate system where the greatest variance lies on the first coordinate (principal component), second greatest variance on the second coordinate, and so on. The document discusses how PCA can be used for face recognition by applying it to image datasets of faces. It reduces the dimensionality of the image data while preserving the key information needed to distinguish different faces. Experimental results show PCA provides reasonably accurate face recognition with low error rates.
Scene Description From Images To SentencesIRJET Journal
This document presents an approach for generating sentences to describe images using distributed intelligence. It involves detecting objects in images using YOLO detection, finding relative positions of objects, labeling background scenes, generating tuples of objects/scenes/relations, extracting candidate sentences from Wikipedia containing tuple elements, searching images for each sentence and selecting the sentence whose images most closely match the input image. The approach is compared to the Babytalk model using BLEU and ROUGE scores, showing comparable performance. Future work to improve object detection and use larger knowledge sources is discussed.
It is a well-known fact that precise definitions play significant role in the development of correct and robust
software. It has been recognized and emphasized that appropriately defined formal conceptual framework
of the context/problem domain proves quite useful in ensuring precise definitions, including those for
software metrics, which are consistent, unambiguous and language independent. In this paper, a formal
conceptual framework for defining metrics for component-based system is proposed, where the framework
formalises the behavioural aspects of the problem domain. The framework in respect of structural aspects
has been discussed in another paper.
Performance Comparision of Machine Learning AlgorithmsDinusha Dilanka
In this paper Compare the performance of two
classification algorithm. I t is useful to differentiate
algorithms based on computational performance rather
than classification accuracy alone. As although
classification accuracy between the algorithms is similar,
computational performance can differ significantly and it
can affect to the final results. So the objective of this paper
is to perform a comparative analysis of two machine
learning algorithms namely, K Nearest neighbor,
classification and Logistic Regression. In this paper it
was considered a large dataset of 7981 data points and 112
features. Then the performance of the above mentioned
machine learning algorithms are examined. In this paper
the processing time and accuracy of the different machine
learning techniques are being estimated by considering the
collected data set, over a 60% for train and remaining
40% for testing. The paper is organized as follows. In
Section I, introduction and background analysis of the
research is included and in section II, problem statement.
In Section III, our application and data analyze Process,
the testing environment, and the Methodology of our
analysis are being described briefly. Section IV comprises
the results of two algorithms. Finally, the paper concludes
with a discussion of future directions for research by
eliminating the problems existing with the current
research methodology.
Foundations for New Champlain Bridge Corridor ProjectDaksh Raj Chopra
In this presentation, we constructed 15 temporary bent foundations for the New Champlain Bridge. The presentation is about the project management done for this project.
This document discusses the Maggi noodles quality management case in India. It describes how Maggi was founded in Switzerland in the late 19th century and was later acquired by Nestle. In 2015, samples of Maggi noodles in India were found to contain illegal levels of lead and MSG contrary to packaging. This led to a nationwide recall of Maggi noodles in India and significant financial losses for Nestle India. After the case, changes were made at Nestle India including appointing an Indian managing director for the first time in 17 years and plans to relaunch Maggi by November 2015 after clearing additional tests.
Internet of things (IoT) Architecture Security AnalysisDaksh Raj Chopra
This Document Briefly summarizes the Security and Privacy Concern Evaluation of Internet of Things (IoT)’s Three Domain Architecture. The Security implementation challenges faced
by IoT devices are addressed along with newly Added Requirement for these devices. The Architecture which we will be using throughout our analysis is explained so as to a novice
user. We will summarize the possible attacks and countermeasures for each and every domain followed by a developer friendly checklist to be followed for security.
Simulation of a Wireless Sub Network using QualNETDaksh Raj Chopra
This report has two scenarios - First one having 2 connections, UDP and TCP. Another scenario has 4 TCP connections having a comparison with and without fading.
This document describes an electronic travel aid device for the blind using ultrasonic sensors to detect obstacles. It consists of an ultrasonic sensor that transmits ultrasound beams to detect objects within 2-3 meters. The distance to objects is categorized into discrete levels of 1, 2, or 3 meters which are indicated to the user via tactile vibrators. The device also detects water pits using audio signals to inform users. It aims to provide mobility information to visually impaired people to help them safely navigate environments.
This document provides a training report for a 6-week internship at Appin Technology Lab on embedded systems. It includes an acknowledgment section thanking guides and mentors. The report then covers the company profile, project undertaken which was a DTMF controlled home automation system, introduction to the project explaining home automation, modular description of the system, design, coding, test cases, industry applications, and future enhancements. It also includes data sheets and sections on DTMF signaling technology.
- Microcontrollers are small computers contained on a single chip that contain a processor core, memory, and input/output interfaces. They are used in automatically controlled embedded systems.
- The AVR is a family of microcontrollers developed by Atmel in the 1990s. It uses RISC architecture and is commonly used in hobbyist and commercial projects due to its low cost and availability.
- Code is burned onto AVR microcontrollers using a software program called Atmel Studio, which allows writing code in C or assembly language. The program is then loaded onto the microcontroller through its pins.
This document describes a DTMF-based home appliance control system that allows wireless control of devices in a home using a user console with numbered keys. The system uses a DTMF decoder, microcontroller, and wireless transmitter and receiver to encode and transmit user selections via DTMF signals and control appliances accordingly. Key components include an AT89S52 microcontroller, MT8870 DTMF decoder, relays, LEDs, buttons, and other basic electronic components. When a key is pressed, the DTMF signal is encoded and transmitted to the receiver, then decoded by the microcontroller to trigger the correct appliance, allowing remote control of devices like lights and appliances from anywhere.
K12 Tableau Tuesday - Algebra Equity and Access in Atlanta Public Schoolsdogden2
Algebra 1 is often described as a “gateway” class, a pivotal moment that can shape the rest of a student’s K–12 education. Early access is key: successfully completing Algebra 1 in middle school allows students to complete advanced math and science coursework in high school, which research shows lead to higher wages and lower rates of unemployment in adulthood.
Learn how The Atlanta Public Schools is using their data to create a more equitable enrollment in middle school Algebra classes.
The ever evoilving world of science /7th class science curiosity /samyans aca...Sandeep Swamy
The Ever-Evolving World of
Science
Welcome to Grade 7 Science4not just a textbook with facts, but an invitation to
question, experiment, and explore the beautiful world we live in. From tiny cells
inside a leaf to the movement of celestial bodies, from household materials to
underground water flows, this journey will challenge your thinking and expand
your knowledge.
Notice something special about this book? The page numbers follow the playful
flight of a butterfly and a soaring paper plane! Just as these objects take flight,
learning soars when curiosity leads the way. Simple observations, like paper
planes, have inspired scientific explorations throughout history.
How to Manage Opening & Closing Controls in Odoo 17 POSCeline George
In Odoo 17 Point of Sale, the opening and closing controls are key for cash management. At the start of a shift, cashiers log in and enter the starting cash amount, marking the beginning of financial tracking. Throughout the shift, every transaction is recorded, creating an audit trail.
A measles outbreak originating in West Texas has been linked to confirmed cases in New Mexico, with additional cases reported in Oklahoma and Kansas. The current case count is 817 from Texas, New Mexico, Oklahoma, and Kansas. 97 individuals have required hospitalization, and 3 deaths, 2 children in Texas and one adult in New Mexico. These fatalities mark the first measles-related deaths in the United States since 2015 and the first pediatric measles death since 2003.
The YSPH Virtual Medical Operations Center Briefs (VMOC) were created as a service-learning project by faculty and graduate students at the Yale School of Public Health in response to the 2010 Haiti Earthquake. Each year, the VMOC Briefs are produced by students enrolled in Environmental Health Science Course 581 - Public Health Emergencies: Disaster Planning and Response. These briefs compile diverse information sources – including status reports, maps, news articles, and web content– into a single, easily digestible document that can be widely shared and used interactively. Key features of this report include:
- Comprehensive Overview: Provides situation updates, maps, relevant news, and web resources.
- Accessibility: Designed for easy reading, wide distribution, and interactive use.
- Collaboration: The “unlocked" format enables other responders to share, copy, and adapt seamlessly. The students learn by doing, quickly discovering how and where to find critical information and presenting it in an easily understood manner.
CURRENT CASE COUNT: 817 (As of 05/3/2025)
• Texas: 688 (+20)(62% of these cases are in Gaines County).
• New Mexico: 67 (+1 )(92.4% of the cases are from Eddy County)
• Oklahoma: 16 (+1)
• Kansas: 46 (32% of the cases are from Gray County)
HOSPITALIZATIONS: 97 (+2)
• Texas: 89 (+2) - This is 13.02% of all TX cases.
• New Mexico: 7 - This is 10.6% of all NM cases.
• Kansas: 1 - This is 2.7% of all KS cases.
DEATHS: 3
• Texas: 2 – This is 0.31% of all cases
• New Mexico: 1 – This is 1.54% of all cases
US NATIONAL CASE COUNT: 967 (Confirmed and suspected):
INTERNATIONAL SPREAD (As of 4/2/2025)
• Mexico – 865 (+58)
‒Chihuahua, Mexico: 844 (+58) cases, 3 hospitalizations, 1 fatality
• Canada: 1531 (+270) (This reflects Ontario's Outbreak, which began 11/24)
‒Ontario, Canada – 1243 (+223) cases, 84 hospitalizations.
• Europe: 6,814
Multi-currency in odoo accounting and Update exchange rates automatically in ...Celine George
Most business transactions use the currencies of several countries for financial operations. For global transactions, multi-currency management is essential for enabling international trade.
How to Customize Your Financial Reports & Tax Reports With Odoo 17 AccountingCeline George
The Accounting module in Odoo 17 is a complete tool designed to manage all financial aspects of a business. Odoo offers a comprehensive set of tools for generating financial and tax reports, which are crucial for managing a company's finances and ensuring compliance with tax regulations.
pulse ppt.pptx Types of pulse , characteristics of pulse , Alteration of pulsesushreesangita003
what is pulse ?
Purpose
physiology and Regulation of pulse
Characteristics of pulse
factors affecting pulse
Sites of pulse
Alteration of pulse
for BSC Nursing 1st semester
for Gnm Nursing 1st year
Students .
vitalsign
INTRO TO STATISTICS
INTRO TO SPSS INTERFACE
CLEANING MULTIPLE CHOICE RESPONSE DATA WITH EXCEL
ANALYZING MULTIPLE CHOICE RESPONSE DATA
INTERPRETATION
Q & A SESSION
PRACTICAL HANDS-ON ACTIVITY
Title: A Quick and Illustrated Guide to APA Style Referencing (7th Edition)
This visual and beginner-friendly guide simplifies the APA referencing style (7th edition) for academic writing. Designed especially for commerce students and research beginners, it includes:
✅ Real examples from original research papers
✅ Color-coded diagrams for clarity
✅ Key rules for in-text citation and reference list formatting
✅ Free citation tools like Mendeley & Zotero explained
Whether you're writing a college assignment, dissertation, or academic article, this guide will help you cite your sources correctly, confidently, and consistent.
Created by: Prof. Ishika Ghosh,
Faculty.
📩 For queries or feedback: [email protected]
As of Mid to April Ending, I am building a new Reiki-Yoga Series. No worries, they are free workshops. So far, I have 3 presentations so its a gradual process. If interested visit: https://www.slideshare.net/YogaPrincess
https://ldmchapels.weebly.com
Blessings and Happy Spring. We are hitting Mid Season.
CBSE - Grade 8 - Science - Chemistry - Metals and Non Metals - WorksheetSritoma Majumder
Introduction
All the materials around us are made up of elements. These elements can be broadly divided into two major groups:
Metals
Non-Metals
Each group has its own unique physical and chemical properties. Let's understand them one by one.
Physical Properties
1. Appearance
Metals: Shiny (lustrous). Example: gold, silver, copper.
Non-metals: Dull appearance (except iodine, which is shiny).
2. Hardness
Metals: Generally hard. Example: iron.
Non-metals: Usually soft (except diamond, a form of carbon, which is very hard).
3. State
Metals: Mostly solids at room temperature (except mercury, which is a liquid).
Non-metals: Can be solids, liquids, or gases. Example: oxygen (gas), bromine (liquid), sulphur (solid).
4. Malleability
Metals: Can be hammered into thin sheets (malleable).
Non-metals: Not malleable. They break when hammered (brittle).
5. Ductility
Metals: Can be drawn into wires (ductile).
Non-metals: Not ductile.
6. Conductivity
Metals: Good conductors of heat and electricity.
Non-metals: Poor conductors (except graphite, which is a good conductor).
7. Sonorous Nature
Metals: Produce a ringing sound when struck.
Non-metals: Do not produce sound.
Chemical Properties
1. Reaction with Oxygen
Metals react with oxygen to form metal oxides.
These metal oxides are usually basic.
Non-metals react with oxygen to form non-metallic oxides.
These oxides are usually acidic.
2. Reaction with Water
Metals:
Some react vigorously (e.g., sodium).
Some react slowly (e.g., iron).
Some do not react at all (e.g., gold, silver).
Non-metals: Generally do not react with water.
3. Reaction with Acids
Metals react with acids to produce salt and hydrogen gas.
Non-metals: Do not react with acids.
4. Reaction with Bases
Some non-metals react with bases to form salts, but this is rare.
Metals generally do not react with bases directly (except amphoteric metals like aluminum and zinc).
Displacement Reaction
More reactive metals can displace less reactive metals from their salt solutions.
Uses of Metals
Iron: Making machines, tools, and buildings.
Aluminum: Used in aircraft, utensils.
Copper: Electrical wires.
Gold and Silver: Jewelry.
Zinc: Coating iron to prevent rusting (galvanization).
Uses of Non-Metals
Oxygen: Breathing.
Nitrogen: Fertilizers.
Chlorine: Water purification.
Carbon: Fuel (coal), steel-making (coke).
Iodine: Medicines.
Alloys
An alloy is a mixture of metals or a metal with a non-metal.
Alloys have improved properties like strength, resistance to rusting.
Geography Sem II Unit 1C Correlation of Geography with other school subjectsProfDrShaikhImran
The correlation of school subjects refers to the interconnectedness and mutual reinforcement between different academic disciplines. This concept highlights how knowledge and skills in one subject can support, enhance, or overlap with learning in another. Recognizing these correlations helps in creating a more holistic and meaningful educational experience.
Exploring Substances:
Acidic, Basic, and
Neutral
Welcome to the fascinating world of acids and bases! Join siblings Ashwin and
Keerthi as they explore the colorful world of substances at their school's
National Science Day fair. Their adventure begins with a mysterious white paper
that reveals hidden messages when sprayed with a special liquid.
In this presentation, we'll discover how different substances can be classified as
acidic, basic, or neutral. We'll explore natural indicators like litmus, red rose
extract, and turmeric that help us identify these substances through color
changes. We'll also learn about neutralization reactions and their applications in
our daily lives.
by sandeep swamy
MATLAB IMPLEMENTATION OF SELF-ORGANIZING MAPS FOR CLUSTERING OF REMOTE SENSING DATA
1. 1
TRAINING REPORT
OF
SIX MONTHS INDUSTRIAL TRAINING,
UNDERTAKEN
AT
DEFENCE RESEARCH AND DEVELOPMENT
ORGANIZATION
IN
DEFENCE TERRAIN RESEARCH LABORATORY
ON
MATLAB IMPLEMENTATION OF SELF-ORGANIZING MAPS FOR
CLUSTERING OF REMOTE SENSING DATA
SUBMITTED IN PARTIAL FULFILLMENT OF THE DEGREE
OF
B.E. (ECE)
Under the Guidance of: Submitted By:
Name: T S RAWAT Name: DAKSH RAJ CHOPRA
Designation: SCIENTIST ‘E’ University ID No.: B120020081
Department: DEFENCE TERRAIN RESEARCH LABORATORY
Chitkara University, Himachal Pradesh, India.
2. 2
ACKNOWLEDGEMENT
I take this opportunity to express my profound gratitude and deep regards to my guide Sc. T S
Rawat for his exemplary guidance, monitoring and constant encouragement throughout the
course of this thesis. The blessing, help and guidance given by him time to time shall carry me
a long way in the journey of life on which I am about to embark.
I also take this opportunity to express a deep sense of gratitude to Sc. Sujata Das, DTRL for
her cordial support, valuable information and guidance, which helped me in completing this
task through various stages.
I am obliged to staff members of DTRL for the valuable information provided by them in their
respective fields. I am grateful for their cooperation during the period of my assignment.
Lastly, I thank almighty, my parents, brother, sisters and friends for their constant
encouragement without which this assignment would not be possible.
3. 3
PREFACE
There is a famous saying “The theory without practical is lame and practical
without theory is blind.”
This project is the result of one semester training at Defence Terrain Research
Laboratory. This internship is an integral part of “Engineering” course and it aims
at providing a first-hand experience of industry to students. The following
practical experience helps the students to view the industrial work and knowledge
closely. I was really fortunate of getting an opportunity to pursue my Summer
Training in reputed, well established, fast growing and professionally managed
organization like Defence Terrain Research Laboratory assigned “MATLAB
IMPLEMENTATION OF SELF-ORGANIZING MAPS FOR
CLUSTERING OF REMOTE SENSING DATA.”
4. 4
CONTENTS
Sr. No. Topic Page No.
1. Project Undertaken 5
2. Introduction 6
3. Requirement Analysis with Illustrations 10
4. SOM Algorithm 16
5. Experiments during training 20
6. Additional feature of SOM 62
7. Future Enhancement 64
8. Bibliography and References 64
5. 5
1. Project Undertaken
My internship was based on a single objective of Self Organizing Maps (SOM). Now first thing
that comes in our mind is that why this topic. Why do we need SOM and what does it do. All
that will be told in this report. There existed a projection known as Sammon Projection which
showed image of data items having same attributes closer to each other and different one having
more distance. But it could not convert high dimensional to low dimensional image. For that
we needed SOM which work on machine learning process from the neurons (attributes)
provided at the input.
The Self-Organizing Map represents a set of high-dimensional data items as a quantized two
dimensional image in an orderly fashion. Every data item is mapped into one point (node) in
the map, and the distances of the items in the map reflect similarities between the items.
SOM used a data mining technique, that it works on the principle of pattern recognize. It uses
the idea of adaptive networks that started the Artificial Intelligence research. This further helps
in the data analysis. For example while studying 17 different elements, we used to differentiate
of the basis of physical properties but SOM learned the values and made a pattern. After that it
did a clustering of that 17 elements.
The Self-Organizing Map (SOM) is a data-analysis method that visualizes similarity relations
in a set of data items. For instance in economy, it has been applied to the comparison of
enterprises at different levels of abstraction, to assess their relative financial conditions, and to
profile their products and customers.
Next comes the industrial applications of this project. In industry, the monitoring of processes,
systems and machineries by the SOM method has been a very important application, and there
the purpose is to describe the masses of different input states by ordered clusters of typical
states. In science and technology at large, there exist unlimited tasks where the research objects
must be classified on the basis of their inherent properties, to mention the classification of
proteins, genetic sequences and galaxies.
Furthermore applications of SOM are:-
1. Statistical methods at large
(a) Exploratory data analysis
(b) Statistical analysis and organization of texts
2. Industrial analyses, control, and telecommunications
3. Biomedical analyses and applications at large
4. Financial applications
6. 6
2. Introduction
2.1 SOM Representation
Only in some special cases the relation of input items with their projection images is
one-to-one in the SOM. More often, especially in industrial and scientific applications,
the mapping is many-to-one: i.e., the projection images on the SOM are local averages
of the input-data distribution, comparable to the k-means averages in classical vector
quantization. In the VQ, the local averages are represented by a finite set of codebook
vectors. The SOM also uses a finite set of “codebook vectors”, called the models, for
the representation of local averages. An input vector is mapped into a particular node
on the SOM array by comparing it with all of the models, and the best-matching model,
called the winner, is identified, like in VQ. The most essential difference with respect
to the k-means clustering, however, is that the models of the SOM also reflect
topographic relations between the projection images which are similar to those of the
source data. So the SOM is actually a data compression method, which represents the
topographic relations of the data space by a finite set of models on a topographic map,
in an orderly fashion.
In the standard-of-living diagram shown in Fig. 2.1, unique input items are mapped on
unique locations on the SOM. However, in the majority of applications, there are
usually many statistically distributed variations of the input items, and the projection
image that is formed on the SOM then represents clusters of the variations. We shall
make this fact more clearly in examples. So, in its genuine form the SOM differs from
all of the other nonlinearly projecting methods, because it usually represents a big data
set by a much smaller number of models, sometimes also called ”weight vectors” (this
latter term comes from the theory of artificial neural networks), arranged as a
rectangular array of nodes. Each model has the same number of parameters as the
number of features in the input items. However, an SOM model may not be a replica
of any input item but only a local average over a subset of items that are most similar
to it. In this sense the SOM works like the k-means clustering algorithm, but in addition,
in a special learning process, the SOM also arranges the k means into a topographic
order according to their similarity relations. The parameters of the models are variable
and they are adjusted by learning such that, in relation to the original items, the
similarity relations of the models finally approximate or represent the similarity
relations of the original items. It is obvious that an insightful view of the complete data
base can then be obtained at one glance.
7. 7
Fig. 2.1. Structured diagram of the data set chosen to describe the standard of living in 126 countries of the world in the year 1992. The
abbreviated country symbols are concentrated onto locations in the (quantized) display computed by the SOM algorithm. The symbols written
in capital letters correspond to those 78 countries for which at least 28 indicators out of 39 were given, and they were used in the real
computation of the SOM. The symbols written in low case letters correspond to countries for which more than 11 indicator values were
missing, and these countries are projected to locations based on the incomplete comparison of their given attributes with those of the 78
countries.
2.2 SOM Display
It shall be emphasized that unlike in the other projective methods, in the SOM the
representations of the items are not moved anywhere in their “topographic” map for their
ordering. Instead, the adjustable parameters of the models are associated with fixed locations
of the map once and for all, namely, with the nodes of a regular, usually two-dimensional array
(Fig. 2). A hexagonal array, like the pixels on a TV screen, provides the best visualization.
Initially the parameters of the models can even have random values. The correct final values
of the models or “weight vectors” will develop gradually by learning. The representations, i.e.,
the models, become more or less exact replica of the input items when their sets of feature
parameters are tuned towards the input items during learning. The SOM algorithm constructs
the models such that: After learning, more similar models will be associated with nodes that
are closer in the array, whereas less similar models will be situated gradually farther away in
the array.
It may be easier to understand the rather involved learning principles and mathematics of the
SOM, if the central idea is first expressed in the following simple illustrative form. Let X denote
a general input item, which is broadcast to all nodes for its concurrent comparison with all of
the models. Every input data item shall select the model that matches best with the input item,
and this model, called the winner, as well as a subset of its spatial neighbors in the array, shall
be modified for better matching. Like in the k-means clustering, the modification is
concentrated on a selected node that contains the winner model. On the other hand, since a
whole spatial neighbourhood around the winner in the array is modified at a time, the degree
of local, differential ordering of the models in this neighbourhood, due to a smoothing action,
8. 8
will be increased. The successive, different inputs cause corrections in different subsets of
models. The local ordering actions will gradually be propagated over the array.
2.3 SOM – A Neural Model
Many principles in computer science have started as models of neural networks. The first
computers were nicknamed “giant brains,” and the electronic logic circuits used in the first
computers, as contrasted with the earlier electromechanical relay-logic (switching) networks,
were essentially nothing but networks of threshold triggers, believed to imitate the alleged
operation of the neural cells.
The first useful neural-network models were adaptive threshold-logic circuits, in which the
signals were weighted by adaptive (”learning”) coefficients. A significant new aspect
introduced in the 1960s was to consider collective effects in distributed adaptive networks,
which materialized in new distributed associative memory models, multilayer signal-
transforming and pattern-classifying circuits, and networks with massive feedbacks and stable
eigenstates, which solved certain optimization problems.
Against this background, the Self-Organizing Map (SOM) introduced around 1981-82 may be
seen as a model of certain cognitive functions, namely, a network model that is able to create
organized representations of sensory experiences, like the brain maps in the cerebral cortex
and other parts of the central nervous system do. In the first place the SOM gave some hints of
how the brain maps could be formed postnatally, without any genetic control. The first
demonstrations of the SOM exemplified the adaptive formation of sensory maps in the brain,
and stipulated what functional properties are most essential to their formation.
In the early 1970s there were big steps made in pattern recognition (PR) techniques. They
continued the idea of adaptive networks that started the “artificial intelligence (AI)” research.
However, after the introduction of large time-shared computer systems, a lot of computer
scientists took a new course in the AI research, developing complex decision rules by ”heuristic
programming”, by which it became possible to implement, e.g., expert systems, computerized
control of large projects, etc. However, these rules were mainly designed manually.
Nonetheless there was a group of computer scientists who were not happy with this approach:
they wanted to continue the original ideas, and to develop computational methods for new
analytical tasks in information science, such as remote sensing, image analysis in medicine,
and speech recognition. This kind of Pattern Recognition was based on mathematical statistics,
and with the advent of new powerful computers, it too could be applied to large and important
problems.
Notwithstanding the connection between AI and PR research broke in the 1970ies, and the AI
and PR conferences and societies started to operate separately. Although the Self-Organizing
Map research was started in the neural-networks context, its applications were actually
developed in experimental pattern-recognition research, which was using real data. It was first
found promising in speech recognition, but very soon numerous other applications were found
in industry, finance, and science. The SOM is nowadays regarded as a general data-analysis
method in a number of fields.
Data mining has a special flavor in data analysis. When a new research topic is started, one
usually has little understanding of the collected data. With time it happens that new, unexpected
results or phenomena will be found. The meaning often given to automated data mining is that
9. 9
the method is able to discover new, unexpected and surprising results. Even in this book I have
tried to collect simple experiments, in which something quite unexpected will show up.
Consider, for instance, Sec. 12 (”The SOM of some metallic elements”) in which we find that
the ferromagnetic metals are mapped to a tight cluster; this result was not expected, but the
data analysis suggested that the nonmagnetic properties of the metals must have a very strong
correlation with the magnetic ones! Or in Sec. 19 (”Two-class separation of mushrooms on the
basis of visible attributes”) we are clustering the mushrooms on the basis of their visible
attributes only, but this clustering results in a dichotomy of edible vs. poisonous species. In
Sec. 13 we are organizing color vectors by the SOM, and if we use a special scaling of the color
components, we obtain a color map that coincides with the chromaticity map of human color
vision, although this result was in no way expected. Accordingly, it may be safe to say that the
SOM is a genuine data mining method, and it will find its most interesting applications in new
areas of science and technology. I may still mention a couple of other works from real life. In
1997, Naim et al. published a work [61] that clustered middle-distance galaxies according to
their morphology, and found a classification that they characterized as a new finding, compared
with the old standard classification performed by Hubble. In Finland, the pulp mill industries
were believed to represent the state of the art, but a group of our laboratory was conducting
very careful studies of the process by the SOM method [1], and the process experts payed
attention to certain instabilities in the flow of pulp through a continuous digester. This
instability was not known before, and there was sufficient reason to change the internal
structures of the digester so that the instability disappeared.
It seems possible that the SOM will open new vistas into the information science. Not only
does it already have numerous spin-offs in applications, but its role in the theory of cognition
is intriguing. However, its mathematics is still in its infancy and offers new problems especially
for mathematical statistics. A lot of high-level research is going on in this area. Maybe it is not
exaggerated to assert that the SOM presents a new information processing paradigm, or at least
a philosophical line in bioinformatics.
To recapitulate, the SOM is a clustering method, but unlike the usual clustering methods, it is
also a topography-preserving nonlinear projection mapping. On the other hand, while the
other nonlinear projection mappings are also topography-preserving mappings, they do not
average data, like the SOM does.
10. 10
3. Requirement Analysis with Illustrations
3.1 Basic Illustration
Now in this training our job was to do the clustering of data using Self Organising Map tool.
So for that we used the MATLAB software in which SOM ToolBox was used. This tool has
various in-build functions for the clustering. But to use a tool we need to first understand the
tool and then give the required input. Let me give an example, a small code, to give you some
idea of how this tool helps in clustering of data in MATLAB.
In this example we have a data of 17 elements having 12 different properties and this whole
data was given to the SOM Tool.
close all
clear all
clc
X = [2.7 23.8 1.34 5105 .7 187 .22 658 92.4 2450 2800 2.72
6.68 10.8 2.7 3400 .79 16 .051 630.5 38.9 1380 300 39.8
10.5 18.8 .99 2700 .8 357 .056 960.5 25 1930 520 1.58
22.42 6.59 .27 4900 5.2 51.5 .032 2454 28 4800 930 5.3
8.64 31.6 2.25 2300 .51 84 .08 320.5 13 767 240 7.25
8.8 12.6 .54 4720 2.08 60 .1 1489 67 3000 1550 6.8
19.3 14.3 .58 2080 .81 268 .031 1063 15.7 2710 420 2.21
8.92 16.2 .73 3900 1.2 338 .0928 1083 50 2360 1110 1.72
11.34 28.9 2.37 1320 .17 30 .03 327 5.9 1750 220 20.7
1.734 25.5 2.95 4600 .45 145 .25 650 46.5 1097 1350 4.3
8.9 12.9 .53 4970 2.03 53 .108 1450 63 3075 1480 7.35
12.16 11.04 .53 3000 1.15 60 .059 1555 36 2200 950 10.75
21.45 15.23 .36 2690 1.6 60 .032 1773 27 4300 600 10.5
7.86 12.3 .59 5100 2.2 50 .11 1530 66 2500 1520 9.9
7.14 17.1 1.69 3700 .43 97 .092 419.5 26.8 907 430 5.95
7.3 27 1.88 2600 .55 57 .05 231.9 14.2 2270 620 11.3
9.8 13.4 2.92 1800 .32 8.6 .029 271.3 14.1 1490 200 118];
for i = 1:12
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [6 6];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
11. 11
trainlen_coarse, 'neigh', neigh);
som_cplane('hexa',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'ASAICCACPMNPPFZSB';
labels2 = 'lbgrdouubgidtenni';
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(17,2);
t=1; % input serial no.
for u=1:17
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,6) + 1;
cv = floor((c-1)/6) + 1;
if mod(cv,2) == 1
shift1 = -.15;
else
shift1 = .35;
end
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch+shift1,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
Now, if 17 different elements are given from the periodic table then it should be arranged in
17 different clusters. But the result gave a surprise to all of us. There were some elements in
single clusters telling us that there was something similar about those elements.
12. 12
Have a look at the output:-
Fig 3.1 This is the cluster formed after the SOM tool worked with the input given to it. Some of the
elements were present in the same clusters due to similar properties.
You can see that that Nickel, Cobalt and Iron are in the same clusters since they have same
ferromagnetic properties. Also Lead and Cadmium are in the same cluster.
3.2 Commands in MATLAB
Now in the above code there are some commands used in MATLAB which help in the
clustering of this data. Few of them are:-
Min - this finds the minimum value in the vector/matrix.
Max - this finds the minimum value in the vector/matrix.
som_lininit - this initializes the map after calculating the Eigen vectors.
som_batchtrain - trains the given som with the given training data.
som_cplane - this visualizes a 2D plane or U matrix.
sum - adds all the array elements.
Mod - gives the remainder.
Floor - rounds off the number to the lower value.
Text - creates text object at the axes.
There is another thing in the tool that is the neighbourhood which we can change according to
our need and data input. SOM works on the basis of neurons. When weights of neurons change
accordingly our output changes. Let me show this with the help of figures:-
13. 13
Fig 3.2 The red dots are the neurons and three black dots on the either sides are the input sequence.
Fig 3.3 When the upper sequence is selected then the neurons move to that side by calculating the
distance and they move to sequence have the shorter distance. And accordingly the weight of neurons
is changed.
14. 14
Fig 3.4 Now the other sequence was selected and the neurons changed it weights.
So every time an input sequence is selected and the weights of neurons gets changed. At the
end we have the final weight of neuron. Now what is this weight of neuron. This weight is the
machine learning process of the machine. We give them data and there is a neuron with them
which learns the pattern and do the clustering accordingly. This is the basic principle of
artificial intelligence.
3.3 SOM Array Size
Maybe it is necessary to state first that the SOM is visualizing the the entire input-data space,
whereupon its density function ought to become clearly visible.
The SOM is a quantizing method. Assume next that we have enough statistical data items to
visualize the clustering structures of the data space with sufficient accuracy. Then it should be
realized that the SOM is a quantizing method, and has a limited spatial resolution to show the
details of the clusters. Sometimes the data set may contain only few clusters, whereupon a
coarse resolution is sufficient. However, if one suspects that there are interesting fine structures
in the data, then a larger array would be needed for sufficient resolution.
Histograms can be displayed on the SOM array. However, it is also necessary to realize that
the SOM can be used to represent a histogram. The number of input data items that is mapped
onto a node is displayed as a shade of gray, or by a pseudo-color. The statistical accuracy of
such a histogram depends on how many input items are mapped per node on the average . A
very coarse rule-of-thumb may be that about 50 input-data items per node on the average
should be sufficient, otherwise the resolution is limited by the sparsity of data. So, in
visualizing clusters, a compromise must be made between resolution and statistical accuracy.
These aspects should be taken into account especially in statistical studies, where only a limited
number of samples are available.
15. 15
Sizing the SOM by a trial-and-error method. It is not possible to estimate or even guess the
exact size of the array beforehand. It must be determined by the trial-and-error method, after
seeing the quality of the first guess. One may have to test several sizes of the SOM to check
that the cluster structures are shown with a sufficient resolution and statistical accuracy.
Typical SOM arrays range from a few dozen to a few hundred nodes. In special problems, such
as the mapping of documents onto the SOM array, even larger maps with, say, thousands of
nodes, are used. The largest map produced by us has been the SOM of seven million patent
abstracts, for which we constructed a one-million-node SOM.
On the other hand, the SOM may be at its best in the visualization of industrial processes, where
unlimited amounts of measurements can be recorded. Then the size of the SOM array is not
limited by the statistical accuracy but by the computational resources, especially if the SOM
has to be constructed periodically in real time, like in the control rooms of factories.
3.4 SOM Array Shape
Because the SOM is trying to represent the distribution of high-dimensional data items by a
two-dimensional projection image, it may be understandable that the scales of the horizontal
and vertical directions of the SOM array should approximately comply with the extensions of
the input-data distribution in the two principal dimensions, namely, those two orthogonal
directions in which the variances of the data are largest. In complete SOM software packages
there is usually an auxiliary function that makes a traditional two-dimensional image of a high-
dimensional distribution, e.g., the Sammon projection (cf., e.g. [39] in our SOM Toolbox
program package. From its main extensions one can estimate visually what the approximate
ratio of the horizontal and vertical sides of the
SOM array should be.
Special shapes of the array. There exist SOMs in which the array has not been selected as a
rectangular sheet. Its topology may resemble, e.g., a cylinder, torus, or a sphere (cf., e.g., [80]).
There also exist special SOMs in which the structure and number of nodes of the array is
determined dynamically, depending on the received data; cf., e.g., [18]. The special topologies,
although requiring more cumbersome displays, may sometimes be justified, e.g., for the
following reasons. 1. The SOM is sometimes used to define the control conditions in industrial
processes or machineries automatically, directly controlling the actuators. A problem may
occur with the boundaries of the SOM sheet: there are distortions and discontinuities, which
affect the control stability. The toroidal topology seems to solve this problem, because there
are then no boundaries in the SOM. A similar effect is obtained by the spherical topology of
the SOM. (Cf. Subsections.4.5 and 5.3, however.) 2. There may exist data, which are cyclic by
their nature. One may think, for example of the application of the SOM in musicology, where
the degrees of the scales repeat by octaves. Either the cylindrical or toroidal topology will then
map the tones cyclically onto the SOM. The dynamical topology, which adjusts itself to
structured data, is very interesting in itself. There is one particular problem, however: one must
be able to define the condition on which a new structure (branching or cutting of the SOM
network) is due. There do not exist universal conditions of this type, and any numerical limit
can only be defined arbitrarily. Accordingly, the generated structure is then not unique. This
same problem is encountered in other neural network models.
16. 16
4. SOM Algorithm
The first SOMs were constructed by a stepwise-recursive learning algorithm, where, at each
step, a selected patch of models in the SOM array was tuned towards the given input item, one
at a time. Consider again Fig. 2. Let the input data items X this time represent a sequence {x(t)}
of real n-dimensional Euclidean vectors x, where t, an integer, signifies a step in the sequence.
Let theMi, being variable, successively attain the values of another sequence {mi(t)} of n-
dimensional real vectors that represent the successively computed approximations of model
mi. Here i is the spatial index of the node with which mi is associated. The original SOM
algorithm assumes that the following process converges and produces the wanted ordered
values for the models:
mi(t + 1) = mi(t) + hci(t)[x(t) −mi(t)] , (1)
where hci(t) is called the neighborhood function. The neighborhood function has the most
central role in self organization. This function resembles the kernel that is applied in usual
smoothing processes. However, in the SOM, the subscript c is the index of a particular node
(winner) in the array, namely, the one with the model mc(t) that has the smallest Euclidean
distance from x(t):
c = argmini {||x(t) −mi(t)||} . (2)
Equations (1) and (2) can be illustrated as defining a recursive step where first the input data
item x(t) defines or selects the best-matching model (winner) according to Eq.(2). Then,
according to Eq.(1), the model at this node as well as at its spatial neighbors in the array are
modified. The modifications always take place in such a direction that the modified models
will match better with the input. The rates of the modifications at different nodes depend on
the mathematical form of the function hci(t). A much-applied choice for the neighborhood
function hci(t) is
hci(t) = α(t) exp[−sqdist(c, i)/2σ2(t)] , (3)
where α(t) < 1 is a monotonically (e.g., hyperbolically, exponentially, or piecewise linearly)
decreasing scalar function of t, sqdist(c, i) is the square of the geometric distance between the
nodes c and i in the array, and σ(t) is another monotonically decreasing function of t,
respectively. The true mathematical form of σ(t) is not crucial, as long as its value is fairly large
in the beginning of the process, say, on the order of 20 per cent of the longer side of the SOM
array, after which it is gradually reduced to a small fraction of it, usually in a few thousand
steps. The topographic order is developed during this period. On the other hand, after this initial
phase of coarse ordering, the final convergence to nearly optimal values of the models takes
place, say, in an order of magnitude more steps, whereupon α(t) attains values on the order of
.01. For a sufficient statistical accuracy, every model must be updated sufficiently often.
However, we must give a warning: the final value of σ shall never go to zero, because otherwise
the process loses its ordering power. It should always remain, say, above half of the array
spacing. In very large SOM arrays, the final value of σ may be on the order of five per cent of
the shorter side of the array. There are also other possible choices for the mathematical form
of hci(t). One of them, the ”bubble” form, is very simple; in it we have hci = 1 up to a
certain radius from the winner, and zero otherwise.
17. 17
4.1 Stable state of the learning process
In the stationary state of learning, every model is the average of input items projected into its
neighborhood, weighted by the neighborhood function.
Assuming that the convergence to some stable state of the SOM is true, we require that the
expectation values of mi(t + 1) and mi(t) for t→∞ must be equal, while hci is nonzero, where
c = c(x(t)) is the index of the winner node for input x(t). In other words we must have
∀i, Et{hci(x(t) −mi(t))} = 0 . (4)
Here Et is the mathematical expectation value operator over t. In the assumed asymptotic state,
for t → ∞, the mi(t) are independent of t and are denoted by m∗ i . If the expectation values
Et(.) are written, for t → ∞, as (1/t)_t(.), we can write
m∗i = _t hci(t)x(t) _t hci(t). (5)
This, however, is still an implicit expression, since c depends on x(t) and the mi, and must be
solved iteratively. Nonetheless, Eq.(5) shall be used for the motivation of the iterative solution
for the mi, known as the batch computation of the SOM (”Batch Map”).
4.2 Initialization of the models
The learning process can be started with random vectors as the initial values of the model
vectors, but learning is sped up significantly, if certain regular initial values are given to the
models.
A special question concerns the selection of the initial values for the mi. It has been
demonstrated by [39] that they can be selected even as random vectors, but a significantly
faster convergence follows if the initial values constitute a regular, two-dimensional sequence
of vectors taken along a hyperplane spanned by the two largest principal components of x (i.e.,
the principal components associated with the two highest eigenvalues); cf. [39]. This method
is called linear initialization.
The initialization of the models as random vectors was originally used only to demonstrate the
capability of the SOM to become ordered, starting from an arbitrary initial state. In practical
applications one expects to achieve the final ordering as quickly as possible, so the selection
of a good initial state may speed up the convergence of the algorithms by orders of magnitude.
4.3 Point density of the models (one-dimensional case)
It was stated in Subsec. 4.1 that in the stationary state of learning, every model vector is the
average of input items projected into its neighborhood, weighted by the neighborhood function.
However, this condition does not yet tell anything about the distribution of the model vectors,
or their point density.
18. 18
To clarify what is thereby meant, we have to revert to the classical vector quantization, or the
k-means algorithm [19], [20], which differs from the SOM in that only the winners are updated
in training; in other words, the ”neighborhood function” hci in k-means learning is equal to
δci, where δci = 1, if c = i, and δci = 0, if c _= i.
No topographic order of the models is produced in the classical vector quantization, but its
mathematical theory is well established. In particular, it has been shown that the point density
q(x) of its model vectors depends on the probability density function p(x) of the input vectors
such that (in the Euclidean metric) q(x) = C · p(x)1/3 , where C is a scalar constant. No similar
result has been derived for general vector dimensionalities in the SOM. In the case that (i) when
the input items are scalar-valued, (ii) when the SOM array is linear, i.e., a one-dimensional
chain, (iii) when the neighborhood function is a box function with N neighbors on each side of
the winner, and (iv) if the SOM contains a very large number of model vectors over a finite
range, Ritter and Schulten [74] have derived the following formula, where C is some constant:
q(x) = C · p(x)r , where
r = 2/3 − 1/(32 + 3(N + 1)2) .
For instance, when N = 1 (one neighbor on each side of the winner), we have r = 0.60.
For Gaussian neighborhood functions, Dersch and Tavan have derived a similar result [15].
In other words, the point density of the models in the display is usually proportional to the
probability density function of the inputs, but not linearly: it is flatter. This phenomenon is not
harmful; as a matter it means that the SOM display is more uniform than it would be if it
represented the exact probability density of the inputs.
4.4 Border effects of the SOM
Another salient effect in the SOM, namely, in the planar sheet topology of the array, is that the
point density of the models at the borders of the array is a little distorted. This effect is
illustrated in the case in which the probability density of input is constant in a certain domain
(support of p(x)) and zero outside it. Consider again first the one-dimensional input and linear
SOM array.
In Fig. 4.1 we show a one-dimensional domain (support) [a, b], over which the probability
density function p(x) of a scalar-valued input variable x is constant. The inputs to the SOM
algorithm are picked up from this support at random. The set of ordered scalar-valued models
μi of the resulting one-dimensional SOM has been rendered on the x axis.
Fig. 4.1 Ordered model values μi over a one-dimensional domain.
19. 19
Fig. 4.2. Converged model values μi over a one-dimensional domain of different lengths.
Numerical values of the μi for different lengths of the SOM array are shown in Fig.4.2. It is
discernible that the μi in the middle of the array are reasonably evenly spaced, but close to the
borders the first distance from the border is bigger, the second spacing is smaller, and the next
spacing is again larger than the average. This effect is explained by Fig. 3: in equilibrium, in
the middle of the array, every μi must coincide with the centroid of set Si, where Si represents
all values of x that will be mapped to node i or its neighbors i − 1 and i + 1. However, at the
borders, all values of x that are mapped to node 1 will be in the interval ranging from a to (μ2
+ μ3)/2, and those values of x that will be mapped to node 2 range from a to (μ3 + μ4)/2. Similar
relations can be shown to hold near the end of the chain for μl−1 and μl. Clearly the definition
of the above intervals is unsymmetric in the middle of the array and at the borders, which makes
the spacings different.
20. 20
5. Experiments during training
5.1 Basic Code of Clustering of Random Colours
Now with the baby steps first a code was tried in which there were 10,000 colour vectors. They
were all randomly generated. After that the input was given to the SOM tool. Now there are
two vector values that we have to give so that the SOM can calculate the distance and keep it
in the required cluster. These are radius_coarse and radius_fine. In radius_coarse, we given the
radius from where it will start to calculate and check the neighbours by similar properties and
put it in the desired cluster. Then we have to give the radius_fine in which it has the final radius
values. So in all we have given the bigger circle to check neighbours and then reduce it to find
neighbours. All this is given by the user according to the needs.
close
clear all
clc
msize = [25 25];
lattice = 'rect';
radius_coarse = [10 7];
radius_fine = [7 5];
trainlen_coarse = 50;
trainlen_fine = 50;
X = rand(10000,3); % random input (training) vectors to the SOM
smI = som_lininit(X,'msize',msize,'lattice',lattice,'shape', ...
'sheet');
smC = som_batchtrain(smI,X,'radius',radius_coarse,'trainlen', ...
trainlen_coarse, 'neigh','gaussian');
sm = som_batchtrain(smC,X,'radius',radius_fine,'trainlen', ...
trainlen_fine, 'neigh','gaussian');
M = sm.codebook;
C = zeros(25,25,3);
for i = 1:25
for j = 1:25
p = 25*(i-1) + j;
C(i,j,1) = M(p,1); C(i,j,2) = M(p,2); C(i,j,3) = M(p,3);
end
end
image(C)
21. 21
Fig 5.1. In this the colours of same shade are near to each other and clusters are made accordingly.
5.2 Hyperspectral Analysis of Soil
In this we had a hyperspectral data of 25 different soils. It had spectral values of different soils
at various wavelengths. We chose three different range of wavelengths – Visible, Infrared,
Short wave Infrared.
Aim: To take a range of spectral values and perform a SOM on it to make a spectral library.
Procedure: For identifying any random value that which soil is that we added noise in the data
of different SNR and check that which noise is giving the result equivalent to the original soil
value.
5.2.1 Soil Type – I Very Dark Grayish Brown Silty Loam
5.2.1.1 Visible Range
The visible spectrum is the portion of the electromagnetic spectrum that is visible to
the human eye. Electromagnetic radiation in this range of wavelengths is called visible light or
simply light. A typical human eye will respond to wavelengths from about 390 to 700 nm. In
terms of frequency, this corresponds to a band in the vicinity of 430–770 THz.
The spectrum does not, however, contain all the colors that the human eyes and brain can
distinguish. Unsaturated colors such as pink, or purple variations such as magenta, are absent,
for example, because they can be made only by a mix of multiple wavelengths. Colors
containing only one wavelength are also called pure colors or spectral colors.
22. 22
Visible wavelengths pass through the "optical window", the region of the electromagnetic
spectrum that allows wavelengths to pass largely unattenuated through the Earth's atmosphere.
An example of this phenomenon is that clean air scatters blue light more than red wavelengths,
and so the midday sky appears blue. The optical window is also referred to as the "visible
window" because it overlaps the human visible response spectrum. The near infrared (NIR)
window lies just out of the human vision, as well as the Medium Wavelength IR (MWIR)
window, and the Long Wavelength or Far Infrared (LWIR or FIR) window, although other
animals may experience them.
Colors that can be produced by visible light of a narrow band of wavelengths (monochromatic
light) are called pure spectral colors. The various color ranges indicated in the illustration are
an approximation: The spectrum is continuous, with no clear boundaries between one color
and the next.
This was a small introduction about Visible range of wavelength. When choosing the visible
range we got a range of spectrum and then we experimented it with addition of white Gaussian
noise in the spectrum.
Now, What is White Gaussian Noise?
Additive white Gaussian noise (AWGN) is a basic noise model used in Information theory to
mimic the effect of many random processes that occur in nature. The modifiers denote specific
characteristics:
Additive because it is added to any noise that might be intrinsic to the information system.
White refers to the idea that it has uniform power across the frequency band for the
information system. It is an analogy to the color white which has uniform emissions at all
frequencies in the visible spectrum.
Gaussian because it has a normal distribution in the time domain with an average time
domain value of zero.
Wideband noise comes from many natural sources, such as the thermal vibrations of atoms in
conductors (referred to as thermal noise or Johnson-Nyquist noise), shot noise,black body
23. 23
radiation from the earth and other warm objects, and from celestial sources such as the Sun.
The central limit theorem of probability theory indicates that the summation of many random
processes will tend to have distribution called Gaussian or Normal.
AWGN is often used as a channel model in which the only impairment to communication is a
linear addition of wideband or white noise with a constant spectral density(expressed
as watts per hertz of bandwidth) and a Gaussian distribution of amplitude. The model does not
account for fading, frequency selectivity, interference, nonlinearity ordispersion. However, it
produces simple and tractable mathematical models which are useful for gaining insight into
the underlying behavior of a system before these other phenomena are considered.
The AWGN channel is a good model for many satellite and deep space communication links.
It is not a good model for most terrestrial links because of multipath, terrain blocking,
interference, etc. However, for terrestrial path modeling, AWGN is commonly used to simulate
background noise of the channel under study, in addition to multipath, terrain blocking,
interference, ground clutter and self interference that modern radio systems encounter in
terrestrial operation.
Code for addition of Additive White Gaussian Noise
1. For SNR 60
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,60)
2. For SNR 58
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,58)
3. For SNR 56
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,56)
24. 24
4. For SNR 54 dB
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,54)
After adding different level of SNR, we get a plot of five different spectral values.
Fig 5.2. Different spectral values of single type of soil (Very dark grayish brown silty loam).
After finding the different spectral values of different SNR, we applied the same clustering
method on it.
Code for clustering
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
ATRANS=textread('soil400clus.txt');
X=ATRANS';
for i = 1:301
mi = min(X(:,i));
25. 25
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
som_cplane('rect',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'SSSSSSS';
labels2 = '1234567';
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(7,2);
t=1; % input serial no.
for u=1:3
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,4) + 1;
cv = floor((c-1)/4) + 1;
if mod(cv,2) == 1
shift1 = -.15;
else
shift1 = .35;
end
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
26. 26
Fig 5.3. Cluster for Soil Type 1
Conclusion: In the above figure, S1 is the original value and rest are the other values with
added noise. We can see that S2 and S5 are near to S1 hence having same properties.
5.2.1.2 Near Infrared
Near-infrared spectroscopy (NIRS) is a spectroscopic method that uses the near-
infrared region of the electromagnetic spectrum (from about 700 nm to 2500 nm). Typical
applications include medical and physiological diagnostics and research including blood
sugar, pulse oximetry, functional neuroimaging, sports medicine, elite sports
training, ergonomics, rehabilitation, neonatal research, brain computer
interface, urology (bladder contraction), and neurology (neurovascular coupling). There are
also applications in other areas as well such as pharmaceutical, food and agrochemical quality
control, atmospheric chemistry, and combustion research.
Near-infrared spectroscopy is based on molecular overtone and combination vibrations. Such
transitions are forbidden by the selection rules of quantum mechanics. As a result, the molar
absorptivity in the near-IR region is typically quite small. One advantage is that NIR can
typically penetrate much farther into a sample than mid infrared radiation. Near-infrared
spectroscopy is, therefore, not a particularly sensitive technique, but it can be very useful in
probing bulk material with little or no sample preparation.
The molecular overtone and combination bands seen in the near-IR are typically very broad,
leading to complex spectra; it can be difficult to assign specific features to specific chemical
components. Multivariate (multiple variables) calibration techniques (e.g., principal
27. 27
components analysis, partial least squares, or artificial neural networks) are often employed to
extract the desired chemical information. Careful development of a set of calibration samples
and application of multivariate calibration techniques is essential for near-infrared analytical
methods.
Code for addition of Additive White Gaussian Noise
1. For SNR 60
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,60)
2. For SNR 58
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,58)
3. For SNR 56
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,56)
4. For SNR 54 dB
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,54)
After adding different level of SNR, we get a plot of five different spectral values.
28. 28
Fig 5.4 Different spectral values of single type of soil (Very dark grayish brown silty loam).
After finding the different spectral values of different SNR, we applied the same clustering
method on it.
Code for clustering
close all
clear all
clc
path(path,'C:UsersDTRLDesktopsoil');
ATRANS=textread('noiseclus700.txt');
X=ATRANS';
for i = 1:251
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
30. 30
Fig 5.5 Clustering for Soil Type I
Conclusion: In the above figure, S1 is the original value and rest are the other values with
added noise. We can see that S2 and S6 are near to S1 hence having same properties.
5.2.1.3 Shortwave Infrared
Short-wave infrared (SWIR) light is typically defined as light in the 0.9 – 1.7μm wavelength
range, but can also be classified from 0.7 – 2.5μm. Since silicon sensors have an upper limit of
approximately 1.0μm, SWIR imaging requires unique opticaland electronic components
capable of performing in the specific SWIR range. Indium gallium arsenide (inGaAs) sensors
are the primary sensors used in SWIR imaging, covering the typical SWIR range, but can
extend as low as 550nm to as high as 2.5μm. Although linear line-scan inGaAs sensors are
commercially available, area-scan inGaAs sensors are typically ITAR restricted. ITAR,
International Treaty and Arms Regulations, is enforced by the government of the United States
of America. ITAR restricted products must adhere to strict export and import regulations for
them to be manufactured and/or sold within and outside of the USA. Nevertheless, lenses such
as SWIR ones can be used for a number of commercial applications with proper licenses.
31. 31
Code for addition of Additive White Gaussian Noise
1. For SNR 60
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,60)
2. For SNR 58
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,58)
3. For SNR 56
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,56)
4. For SNR 54 dB
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,54)
After adding different level of SNR, we get a plot of five different spectral values.
32. 32
Fig 5.6 Different spectral values of single type of soil (Very dark grayish brown silty loam).
After finding the different spectral values of different SNR, we applied the same clustering
method on it.
Code for clustering
close all
clear all
clc
path(path,'C:UsersDTRLDesktopsoil2');
ATRANS=textread('soil1400clus.txt');
X=ATRANS';
for i = 1:424
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
34. 34
Fig 5.7. Cluster for Soil Type I
Conclusion: In the above figure, S1 is the original value and rest are the other values with
added noise. We can see that S2 and S3 are near to S1 hence having same properties.
5.2.2 Soil Type – II Grayish Brown Loam
5.2.2.1 Visible Range
Code for addition of Additive White Gaussian Noise
1. For SNR 60
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,60)
2. For SNR 58
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,58)
35. 35
3. For SNR 56
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,56)
4. For SNR 54 dB
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,54)
After adding different level of SNR, we get a plot of five different spectral values.
Fig 5.8. Different spectral values of single type of soil (Grayish brown loam).
36. 36
After finding the different spectral values of different SNR, we applied the same clustering
method on it.
Code for clustering
close all
clear all
clc
path(path,'C:UsersDTRLDesktopsoil2');
ATRANS=textread('soil400clus.txt');
X=ATRANS';
for i = 1:301
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
som_cplane('rect',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'SSSSSSS';
labels2 = '1234567';
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(7,2);
t=1; % input serial no.
for u=1:7
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,4) + 1;
cv = floor((c-1)/4) + 1;
if mod(cv,2) == 1
shift1 = -.15;
else
shift1 = .35;
end
37. 37
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
op_mat
Fig 5.9. Clustering for Soil Type II
Conclusion: In the above figure, S1 is the original value and rest are the other values with
added noise. We can see that none is near to S1 hence all are different.
38. 38
5.2.2.2 Near Infrared
Code for addition of Additive White Gaussian Noise
1. For SNR 60
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,60)
2. For SNR 58
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,58)
3. For SNR 56
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,56)
4. For SNR 54 dB
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,54)
39. 39
After adding different level of SNR, we get a plot of five different spectral values.
Fig 5.10. Different spectral values of single type of soil (Grayish brown loam).
After finding the different spectral values of different SNR, we applied the same clustering
method on it.
Code for clustering
close all
clear all
clc
path(path,'C:UsersDTRLDesktopsoil2');
ATRANS=textread('soil700clus.txt');
X=ATRANS';
for i = 1:185
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
41. 41
Fig 5.11. Clustering for Soil Type II
Conclusion: In the above figure, S1 is the original value and rest are the other values with
added noise. We can see that S2 is near to S1 hence having same properties.
5.2.2.3 Shortwave Infrared
Code for addition of Additive White Gaussian Noise
1. For SNR 60
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,60)
2. For SNR 58
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,58)
3. For SNR 56
close all
clear all
clc
42. 42
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,56)
4. For SNR 54 dB
close all
clear all
clc
path(path,'F:TraineeDakshsoil3');
A=textread('soil400.txt');
noise=awgn(A,54)
After adding different level of SNR, we get a plot of five different spectral values.
Fig 5.12. Different spectral values of single type of soil (Grayish brown loam).
After finding the different spectral values of different SNR, we applied the same clustering
method on it.
Code for clustering
close all
clear all
clc
path(path,'C:UsersDTRLDesktopsoil2');
ATRANS=textread('soil700clus.txt');
X=ATRANS';
for i = 1:424
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
43. 43
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
som_cplane('rect',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'SSSSSSS';
labels2 = '1234567';
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(7,2);
t=1; % input serial no.
for u=1:5
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,4) + 1;
cv = floor((c-1)/4) + 1;
if mod(cv,2) == 1
shift1 = -.15;
else
shift1 = .35;
end
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
op_mat
44. 44
Fig 5.13. Clustering for Soil Type II
Conclusion: In the above figure, S1 is the original value and rest are the other values with
added noise. We can see that none is near to S1.
Similar experiment was done with other 22 soils.
5.3 Polarimetry Analysis of Different Regions
Now in this experiment we had different types of terrain like Urban, Water and Vegetated. So
we decided to check the different terrain by calculating there Co and Cross Polarization.
Aim: First we plotted the 3D curve for all the values of co and cross in all the different regions
and see the changes in the plots. Then we did the SOM clustering and see if these values it also
tells the same thing.
Procedure: For the clustering we found the wavelets from the array with the help of Haar
wavelet at different levels. Coefficients of wavelets were found and with them the input was
given to the SOM for clustering.
Theory:
Cross – Polarization
Cross polarized wave (XPW) generation is a nonlinear optical process that can be classified
in the group of frequency degenerate [four wave mixing] processes. It can take place only in
media with anisotropy of third order nonlinearity. As a result of such nonlinear optical
interaction at the output of the nonlinear crystal it is generated a new linearly polarized wave
45. 45
at the same frequency, but with polarization oriented perpendicularly to the polarization of
input wave
.
Fig 5.14.
Simplified optical scheme for XPW generation is shown on Fig. 5.14. It consists of a nonlinear
crystal plate (thick 1-2 mm) sandwiched between two crossed polarizers. The intensity of
generated XPW has cubic dependence with respect to the intensity of the input wave. In fact
this is the main reason this effect is so successful for improvement of the contrast of the
temporal and spatial profiles of femtosecond pulses. Since cubic crystals are used as nonlinear
media they are isotropic with respect to linear properties (there is no birefringence) and because
of that the phase and group velocities of both waves XPW and the fundamental wave(FW) are
equal:VXPW=VFW and Vgr,XPW=Vgr,FW. Consequence of that is ideal phase and group velocity
matching for the two waves propagating along the crystal. This property allows obtaining very
good efficiency of the XPW generation process with minimum distortions of the pulse shape
and the spectrum.
Wavelets
A wavelet is a wave-like oscillation with an amplitude that begins at zero, increases, and then
decreases back to zero. It can typically be visualized as a "brief oscillation" like one might see
recorded by a seismograph or heart monitor. Generally, wavelets are purposefully crafted to
have specific properties that make them useful for signal processing. Wavelets can be
combined, using a "reverse, shift, multiply and integrate" technique called convolution, with
portions of a known signal to extract information from the unknown signal.
For example, a wavelet could be created to have a frequency of Middle C and a short duration
of roughly a 32nd note. If this wavelet was to be convolved with a signal created from the
recording of a song, then the resulting signal would be useful for determining when the Middle
C note was being played in the song. Mathematically, the wavelet will correlate with the signal
if the unknown signal contains information of similar frequency. This concept of correlation is
at the core of many practical applications of wavelet theory.
46. 46
As a mathematical tool, wavelets can be used to extract information from many different kinds
of data, including – but certainly not limited to – audio signals and images. Sets of wavelets
are generally needed to analyze data fully. A set of "complementary" wavelets will decompose
data without gaps or overlap so that the decomposition process is mathematically reversible.
Thus, sets of complementary wavelets are useful in wavelet based compression/decompression
algorithms where it is desirable to recover the original information with minimal loss.
In formal terms, this representation is a wavelet series representation of a square-integrable
function with respect to either a complete,orthonormal set of basis functions, or
an overcomplete set or frame of a vector space, for the Hilbert space of square integrable
functions.
Haar Wavelet
In mathematics, the Haar wavelet is a sequence of rescaled "square-shaped" functions which
together form a wavelet family or basis. Wavelet analysis is similar to Fourier analysis in that
it allows a target function over an interval to be represented in terms of anorthonormal basis.
The Haar sequence is now recognised as the first known wavelet basis and extensively used
as a teaching example.
Fig 5.15. Haar Wavelet
47. 47
The Haar sequence was proposed in 1909 by Alfréd Haar.[1]
Haar used these functions to
give an example of an orthonormal system for the space of square-integrable functions on
the unit interval [0, 1]. The study of wavelets, and even the term "wavelet", did not come
until much later. As a special case of the Daubechies wavelet, the Haar wavelet is also known
as Db1.
The Haar wavelet is also the simplest possible wavelet. The technical disadvantage of the
Haar wavelet is that it is not continuous, and therefore not differentiable. This property can,
however, be an advantage for the analysis of signals with sudden transitions, such as
monitoring of tool failure in machines.[2]
The Haar wavelet's mother wavelet function can be described as
Its scaling function can be described as
5.3.1 Urban Region
First we plotted the 3D curve for all the values of co and cross in the urban region and see the
changes in the plots.
Code for 3D Plot
close all
clear all
clc
path(path,'F:Traineeurbanroidata');
X=textread('urbanroi1_co.txt');
path(path,'F:Traineeurbanroidata');
A=textread('urbanroi2_co.txt');
path(path,'F:Traineeurbanroidata');
B=textread('urbanroi3_co.txt');
path(path,'F:Traineeurbanroidata');
C=textread('urbanroi4_co.txt');
path(path,'F:Traineeurbanroidata');
D=textread('urbanroi5_co.txt');
path(path,'F:Traineeurbanroidata');
E=textread('urbanroi1_cross.txt');
path(path,'F:Traineeurbanroidata');
F=textread('urbanroi2_cross.txt');
path(path,'F:Traineeurbanroidata');
50. 50
Fig 5.16. 3D Plot
Code for finding the Wavelet Coefficient (Level 2)
clear all
clc
close all
% The current extension mode is zero-padding (see dwtmode).
% Load original image.
% X contains the loaded image.
% Perform decomposition at level 2
% of X using db1.
path(path,'F:Trainee');
X=textread('urbanroi3_cross.txt');
[c,s] = wavedec2(X,2,'haar');
sizex = size(X)
sizec = size(c)
val_s = s
% Extract details coefficients at level 2
% in each orientation, from wavelet decomposition
% structure [c,s].
D = detcoef2('c',c,s,2)
51. 51
Code for clustering data at Level 2
close all
clear all
clc
path(path,'F:Traineeurbanroidatahaar');
X=textread('totaln3.txt');
for i = 1:5
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
som_cplane('rect',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'CCCCC';
labels2 = '12345';
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(5,2);
t=1; % input serial no.
for u=1:6348
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,4) + 1;
cv = floor((c-1)/4) + 1;
if mod(cv,2) == 1
shift1 = -.15;
else
shift1 = .35;
end
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
52. 52
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
op_mat
Fig 5.17. Clustering at Level 2
Conclusion: In the above figure, all are different classes (co and cross) derived from the
wavelet coefficients. Now we can see that C3 and C4 are near to each other hence having same
properties.
Code for finding the Wavelet Coefficient (Level 3)
clear all
clc
close all
% The current extension mode is zero-padding (see dwtmode).
% Load original image.
% X contains the loaded image.
% Perform decomposition at level 3
% of X using db1.
path(path,'F:Traineeurbanroidata');
X=textread('urbanroi4_co.txt');
[c,s] = wavedec2(X,3,'haar');
53. 53
sizex = size(X)
sizec = size(c)
val_s = s
% Extract details coefficients at level 3
% in each orientation, from wavelet decomposition
% structure [c,s].
D = detcoef2('c',c,s,3)
Code for clustering data at Level 3
close all
clear all
clc
path(path,'F:Traineeurbanroidatahaar');
X=textread('totaln3.txt');
for i = 1:5
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
som_cplane('rect',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'CCCCC';
labels2 = '12345';
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(5,2);
t=1; % input serial no.
for u=1:1656
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,4) + 1;
cv = floor((c-1)/4) + 1;
54. 54
if mod(cv,2) == 1
shift1 = -.15;
else
shift1 = .35;
end
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
op_mat
Fig 5.18. Clustering at Level 3
Conclusion: In the above figure, all are different classes (co and cross) derived from the
wavelet coefficients. Now we can see that C3 and C4 are near to each other hence having same
properties.
55. 55
Code for finding the Wavelet Coefficient (Level 4)
clear all
clc
close all
% The current extension mode is zero-padding (see dwtmode).
% Load original image.
% X contains the loaded image.
% Perform decomposition at level 4
% of X using db1.
path(path,'F:Traineeurbanroidata');
X=textread('urbanroi3_cross.txt');
[c,s] = wavedec2(X,4,'haar');
sizex = size(X)
sizec = size(c)
val_s = s
% Extract details coefficients at level 4
% in each orientation, from wavelet decomposition
% structure [c,s].
D = detcoef2('c',c,s,4)
Code for clustering data at Level 4
close all
clear all
clc
path(path,'F:Traineeurbanroidatahaar');
X=textread('totaln4.txt');
for i = 1:5
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
som_cplane('rect',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'CCCCC';
labels2 = '12345';
56. 56
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(5,2);
t=1; % input serial no.
for u=1:432
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,4) + 1;
cv = floor((c-1)/4) + 1;
if mod(cv,2) == 1
shift1 = -.15;
else
shift1 = .35;
end
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
op_mat
57. 57
Fig 5.19. Clustering at Level 4
Conclusion: In the above figure, all are different classes (co and cross) derived from the
wavelet coefficients. Now we can see that C1 and C5 are near to each other hence having same
properties.
Code for finding the Wavelet Coefficient (Level 5)
clear all
clc
close all
% The current extension mode is zero-padding (see dwtmode).
% Load original image.
% X contains the loaded image.
% Perform decomposition at level 5
% of X using db1.
path(path,'F:Trainee');
X=textread('urbanroi5_co.txt');
[c,s] = wavedec2(X,5,'haar');
sizex = size(X)
sizec = size(c)
val_s = s
% Extract details coefficients at level 5
% in each orientation, from wavelet decomposition
% structure [c,s].
D = detcoef2('c',c,s,5)
58. 58
Code for clustering data at Level 5
close all
clear all
clc
path(path,'F:Traineeurbanroidatahaar');
X=textread('totaln5.txt');
for i = 1:5
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
som_cplane('rect',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'CCCCC';
labels2 = '12345';
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(5,2);
t=1; % input serial no.
for u=1:108
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,4) + 1;
cv = floor((c-1)/4) + 1;
if mod(cv,2) == 1
shift1 = -.15;
else
shift1 = .35;
end
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
59. 59
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
op_mat
Fig 5.20. Clustering at Level 5
Conclusion: In the above figure, all are different classes (co and cross) derived from the
wavelet coefficients. Now we can see that C1 and C4 are near to each other hence having same
properties.
Code for finding the Wavelet Coefficient (Level 6)
clear all
clc
close all
% The current extension mode is zero-padding (see dwtmode).
% Load original image.
% X contains the loaded image.
% Perform decomposition at level 6
% of X using db1.
path(path,'F:Trainee');
X=textread('urbanroi5_cross.txt');
[c,s] = wavedec2(X,6,'haar');
60. 60
sizex = size(X)
sizec = size(c)
val_s = s
% Extract details coefficients at level 6
% in each orientation, from wavelet decomposition
% structure [c,s].
D = detcoef2('c',c,s,6)
Code for clustering data at Level 6
close all
clear all
clc
path(path,'F:Traineeurbanroidatahaar');
X=textread('totaln6.txt');
for i = 1:5
mi = min(X(:,i));
ma = max(X(:,i));
X(:,i) = (X(:,i)-mi)/(ma - mi);
end
msize = [4 4];
lattice = 'hexa'; % hexagonal lattice
neigh = 'gaussian'; % neighborhood function
radius_coarse = [4 .5]; % [initial final]
trainlen_coarse = 50; % cycles in coarse training
radius_fine = [.5 .5]; % [initial final]
trainlen_fine = 10; % cycles in fine training
smI = som_lininit(X, 'msize', msize, 'lattice', lattice, 'shape', ...
'sheet');
smC = som_batchtrain(smI, X, 'radius', radius_coarse, 'trainlen', ...
trainlen_coarse, 'neigh', neigh);
som_cplane('rect',msize, 'none')
sm = som_batchtrain(smC, X, 'radius', radius_fine, 'trainlen', ...
trainlen_fine, 'neigh', neigh);
labels1 = 'CCCCC';
labels2 = '12345';
M = sm.codebook;
norms2 = sum(M.*M,2);
op_mat=zeros(5,2);
t=1; % input serial no.
for u=1:36
X1 = X(u,:)';
Y = norms2 - 2*M*X1;
[C,c] = min(Y);
op_mat(u,1)=t;
op_mat(u,2)=c;
ch = mod(c-1,4) + 1;
cv = floor((c-1)/4) + 1;
if mod(cv,2) == 1
61. 61
shift1 = -.15;
else
shift1 = .35;
end
if u==9 || u==11
shift2 = -.3;
elseif u==5 || u==14
shift2 = .3;
else
shift2 = 0;
end
% text(ch+shift1,cv+shift2,[labels1(u) ...
% labels2(u)], 'FontSize',15);
text(ch,cv+shift2,[labels1(u) ...
labels2(u)], 'FontSize',15);
t=t+1;
end
op_mat
Fig 5.21. Clustering at Level 6
Conclusion: In the above figure, all are different classes (co and cross) derived from the
wavelet coefficients. Now we can see that C3 and C5 are near to each other hence having same
properties.
62. 62
6. Additional Feature of SOM
A graphic display based on these distances, called the U matrix, is explained in this section.
The clustering tendency of the data, or the models to describe them, can be shown graphically
based on the magnitudes of vectorial distances between neighboring model s in the map, as
shown in Fig. 6.1. below. In it, the number of hexagonal cells has first been increased from 18
by 17 to 35 by 33, in order to create blank interstitial cells that can be colored by shades of
gray or by pseudocolors to emphasize the cluster borders. This creation of the interstitial cells
is made automatically, when the instructions defined below are executed.
The U matrix. A graphic display called theU matrix has been developed by Ultsch [87], as
well as Kraaijveld et al. [47], to illustrate the degree of clustering tendency on the SOM. In the
basic method, interstitial (e.g. hexagonal in the present example) cells are added between the
original SOM cells in the display. So, if we have an SOM array of 18 by 17 cells, after addition
of the new cells the array size becomes 35 by 33. Notice, however, that the extra cells are not
involved in the SOM algorithm, only the original 18 by 17 cells were trained. The average
(smoothed) distances between the nearest SOM models are represented by light colors for small
mean differences, and darker colors for larger mean differences. A ”cluster landscape,” formed
over the SOM, then visualizes the degree of classification. The groundwork for the U matrix is
generated by the instructions
colormapigray = ones(64,3) - colormap(’gray’);
colormap(colormapigray);
msize = [18 17];
Um = som_umat(sm);
som_cplane(’hexaU’, sm.topol.msize, Um(:));
In this case, we need not draw the blank SOM groundwork by the instruction som
cplane(’hexa’,msize, ’none’) . The example with the countries in August 2014 will now be
represented together with the U-matrix ”landscape” in Fig. 6.1.. It shows darker “ravines”
between the clusters. Annotation of the SOM nodes. After that, the country acronyms are
written on it by the text instruction. The complete map is shown in Fig. 6.1.
63. 63
Fig. 6.1. The SOM, with the U matrix, of the financial status of 50 countries
A particular remark is due here. Notice that there are plenty of unlabelled areas in the SOM.
During training, when the model vectors had not yet reached their final values, the ”winner
nodes” for certain countries were located in completely different places than finally;
nonetheless the models at these nodes gathered memory traces during training, too. Thus, these
nodes have learned more or less wrong and even random values during the coarse training,
with the result that the vectorial differences of the models in those places are large. So the SOM
with unique items mapped on it and having plenty of blank space between them is not
particularly suitable for the demonstration of the U matrix, although some interesting details
can be found in it.
64. 64
7. Future Enhancements
Artificial neural Networks is currently a hot research area in Data compression. GSOM
algorithm for data compression being a wide field which is rapidly finding use in many applied
fields and technologies GSOM has some limitation. They cannot compress higher size of audio
and video file. So To Improve the Compression Ratio of higher size of audio and video file in
future enhancement.
8. Bibliography and References
1. IJCSNS International Journal of Computer Science and Network Security, VOL.14
No.11, November 2014
2. https://en.wikipedia.org/wiki/Haar_wavelet
3. https://en.wikipedia.org/wiki/Wavelet
4. https://en.wikipedia.org/wiki/Cross-polarized_wave_generation
5. http://www.edmundoptics.com/resources/application-notes/imaging/what-is-swir/
6. https://en.wikipedia.org/wiki/Near-infrared_spectroscopy
7. https://en.wikipedia.org/wiki/Additive_white_Gaussian_noise
8. https://en.wikipedia.org/wiki/Visible_spectrum
9. https://en.wikipedia.org/wiki/Visible_spectrum#/media/File:Linear_visible_spectrum.
svg
10. Kohonen, T., MATLAB Implementations and Applications of the Self-Organizing
Map, Unigrafia Oy, Helsinki, Finland, 2014.
11. Introduction to Kohanen’s Self Organising Maps, Fernando Bacao and Victor Lobo