This lesson covers the core data science related content required for applying ML. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Dimensionality reduction techniques are used to reduce the number of features or variables in a dataset. This helps simplify models and improve performance. Principal component analysis (PCA) is a common technique that transforms correlated variables into linearly uncorrelated principal components. Other techniques include backward elimination, forward selection, filtering out low variance or highly correlated features. Dimensionality reduction benefits include reducing storage space, faster training times, and better visualization of data.
Statistical theory is a branch of mathematics and statistics that provides the foundation for understanding and working with data, making inferences, and drawing conclusions from observed phenomena. It encompasses a wide range of concepts, principles, and techniques for analyzing and interpreting data in a systematic and rigorous manner. Statistical theory is fundamental to various fields, including science, social science, economics, engineering, and more.
Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...Maninda Edirisooriya
Bias and Variance are the deepest concepts in ML which drives the decision making of a ML project. Regularization is a solution for the high variance problem. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Dimensionality reduction techniques transform high-dimensional data into a lower-dimensional representation while retaining important information. Principal component analysis (PCA) is a common linear technique that projects data along directions of maximum variance to obtain principal components as new uncorrelated variables. It works by computing the covariance matrix of standardized data to identify correlations, then computes the eigenvalues and eigenvectors of the covariance matrix to identify the principal components that capture the most information with fewer dimensions.
Data preprocessing is the process of preparing raw data for analysis by cleaning it, transforming it, and reducing it. The key steps in data preprocessing include data cleaning to handle missing values, outliers, and noise; data transformation techniques like normalization, discretization, and feature extraction; and data reduction methods like dimensionality reduction and sampling. Preprocessing ensures the data is consistent, accurate and suitable for building machine learning models.
data reduction techniques-data minig.pptxfarheengul004
this slide covers all data reduction techniques of data mining with descriptive examples and all further algorithms. the prerequisite is machine learning.
This document discusses feature engineering, which is the process of transforming raw data into features that better represent the underlying problem for predictive models. It covers feature engineering categories like feature selection, feature transformation, and feature extraction. Specific techniques covered include imputation, handling outliers, binning, log transforms, scaling, and feature subset selection methods like filter, wrapper, and embedded methods. The goal of feature engineering is to improve machine learning model performance by preparing proper input data compatible with algorithm requirements.
Dimensionality Reduction in Machine LearningRomiRoy4
This document discusses dimensionality reduction techniques. Dimensionality reduction reduces the number of random variables under consideration to address issues like sparsity and less similarity between data points. It is accomplished through feature selection, which omits redundant/irrelevant features, or feature extraction, which maps features into a lower dimensional space. Dimensionality reduction provides advantages like less complexity, storage needs, computation time and improved model accuracy. Popular techniques include principal component analysis (PCA), which extracts new variables, and filtering methods. PCA involves standardizing data, computing correlations via the covariance matrix, and identifying principal components via eigenvectors and eigenvalues.
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
This document provides an introduction to the CSC112 Algorithms and Data Structures lecture. It discusses the need for data structures to organize data efficiently and enable more complex applications. Different types of data structures are presented, including linear structures like arrays, lists, queues and stacks, as well as non-linear structures like trees and graphs. Key data structure operations like traversing, searching, inserting and deleting records are also outlined. The document emphasizes that the choice of data structure and algorithm can significantly impact a program's efficiency and performance.
The document discusses feature engineering for machine learning. It defines feature engineering as the process of transforming raw data into features that better represent the data and improve machine learning performance. Some key techniques discussed include feature selection, construction, transformation, and extraction. Feature construction involves generating new features from existing ones, such as calculating apartment area from length and breadth. Feature extraction techniques discussed are principal component analysis, which transforms correlated features into linearly uncorrelated components capturing maximum variance. The document provides examples and steps for principal component analysis.
MACHINE LEARNING YEAR DL SECOND PART.pptxNAGARAJANS68
The document discusses various concepts related to machine learning models including prediction errors, overfitting, underfitting, bias, variance, hyperparameter tuning, and regularization techniques. It provides explanations of key terms and challenges in machine learning like the curse of dimensionality. Cross-validation methods like k-fold are presented as ways to evaluate model performance on unseen data. Optimization algorithms such as gradient descent and stochastic gradient descent are covered. Regularization techniques like Lasso, Ridge, and Elastic Net are introduced.
This document summarizes a proposed method for discriminative unsupervised dimensionality reduction called DUDR. It begins by introducing traditional dimensionality reduction techniques like PCA and LDA. It then discusses limitations of existing graph embedding methods that require constructing a graph beforehand. The proposed DUDR method jointly learns the graph construction and dimensionality reduction to avoid this dependency. It formulates an optimization problem to learn a projection matrix and affinity matrix simultaneously. Experimental results on synthetic and real-world datasets show DUDR achieves better clustering performance than other methods like PCA, LPP, k-means and NMF.
Principal Component Analysis (PCA) is an unsupervised learning algorithm used for dimensionality reduction. It transforms correlated variables into linearly uncorrelated variables called principal components. PCA works by considering the variance of each attribute to reduce dimensionality while preserving as much information as possible. It is commonly used for exploratory data analysis, predictive modeling, and visualization.
PCA is a dimensionality reduction technique that transforms a number of potentially correlated variables into a smaller number of uncorrelated variables called principal components. It works by geometrically projecting the data onto a lower dimensional space in a way that retains as much of the variation present in the original data as possible. PCA is widely used for applications like data visualization, feature extraction and dimensionality reduction. An example of using PCA in genetics is shown, where the variability in human DNA data can be strongly linked to geographical patterns through principal component analysis.
Feature extraction and selection are important techniques in machine learning. Feature extraction transforms raw data into meaningful features that better represent the data. This reduces dimensionality and complexity. Good features are unique to an object and prevalent across many data samples. Principal component analysis is an important dimensionality reduction technique that transforms correlated features into linearly uncorrelated principal components. This both reduces dimensionality and preserves information.
This document discusses the importance of data preprocessing techniques for improving data quality. It outlines several key steps in data preprocessing: data cleaning, which detects and corrects errors and inconsistencies; data integration, which merges data from multiple sources; data reduction, which reduces data dimensions or volumes; and data transformation, which converts data into appropriate forms for analysis. Specific techniques discussed include missing value imputation, binning, smoothing, normalization, and discretization. The overall goal of data preprocessing is to prepare raw data for data mining and ensure high quality results.
Machine learning lets you make better business decisions by uncovering patterns in your consumer behavior data that is hard for the human eye to spot. You can also use it to automate routine, expensive human tasks that were previously not doable by computers. In the business to business space (B2B), if your competitors can make wiser business decisions based on data and automate more business operations but you still base your decisions on guesswork and lack automation, you will lose out on business productivity. In this introduction to machine learning tech talk, you will learn how to use machine learning even if you do not have deep technical expertise on this technology.
Topics covered:
1.What is machine learning
2.What is a typical ML application architecture
3.How to start ML development with free resource links
4.Key decision factors in ML technology selection depending on use case scenarios
Feature engineering is the process of selecting, modifying, or creating new features (variables) from raw data to improve the performance of machine learning models. It involves identifying the most relevant features, transforming data into a suitable format, handling missing values, encoding categorical variables, scaling numerical data, and creating interaction terms or derived features. Effective feature engineering can significantly enhance a model's accuracy and interpretability by providing it with the most informative inputs. It is often considered a crucial step in the machine learning pipeline.
Introduction to Data Structures Sorting and searchingMvenkatarao
This document provides an overview of data structures and algorithms. It begins by defining a data structure as a way of storing and organizing data in a computer so that it can be used efficiently by algorithms. Data structures can be primitive, directly operated on by machine instructions, or non-primitive, developed from primitive structures. Linear structures maintain adjacency between elements while non-linear do not. Common operations on data structures include adding, deleting, traversing, sorting, searching, and updating elements. The document also defines algorithms and their properties, including finiteness, definiteness, inputs, outputs, and effectiveness. It discusses analyzing algorithms based on time and space complexity and provides examples of different complexities including constant, logarithmic, linear, quadratic,
Lecture - 10 Transformer Model, Motivation to Transformers, Principles, and ...Maninda Edirisooriya
Learn about the limitations of earlier Deep Sequence Models like RNNs, GRUs and LSTMs; Evolution of Attention Model as the Transformer Model with the paper, "Attention is All You Need". This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2024 first half of the year.
Learn End-to-End Learning, Multi-Task Learning, Transfer Learning and Meta Learning. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2024 first half of the year.
More Related Content
Similar to Lecture 8 - Feature Engineering and Optimization, a lecture in subject module Statistical & Machine Learning (20)
This document discusses feature engineering, which is the process of transforming raw data into features that better represent the underlying problem for predictive models. It covers feature engineering categories like feature selection, feature transformation, and feature extraction. Specific techniques covered include imputation, handling outliers, binning, log transforms, scaling, and feature subset selection methods like filter, wrapper, and embedded methods. The goal of feature engineering is to improve machine learning model performance by preparing proper input data compatible with algorithm requirements.
Dimensionality Reduction in Machine LearningRomiRoy4
This document discusses dimensionality reduction techniques. Dimensionality reduction reduces the number of random variables under consideration to address issues like sparsity and less similarity between data points. It is accomplished through feature selection, which omits redundant/irrelevant features, or feature extraction, which maps features into a lower dimensional space. Dimensionality reduction provides advantages like less complexity, storage needs, computation time and improved model accuracy. Popular techniques include principal component analysis (PCA), which extracts new variables, and filtering methods. PCA involves standardizing data, computing correlations via the covariance matrix, and identifying principal components via eigenvectors and eigenvalues.
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
This document provides an introduction to the CSC112 Algorithms and Data Structures lecture. It discusses the need for data structures to organize data efficiently and enable more complex applications. Different types of data structures are presented, including linear structures like arrays, lists, queues and stacks, as well as non-linear structures like trees and graphs. Key data structure operations like traversing, searching, inserting and deleting records are also outlined. The document emphasizes that the choice of data structure and algorithm can significantly impact a program's efficiency and performance.
The document discusses feature engineering for machine learning. It defines feature engineering as the process of transforming raw data into features that better represent the data and improve machine learning performance. Some key techniques discussed include feature selection, construction, transformation, and extraction. Feature construction involves generating new features from existing ones, such as calculating apartment area from length and breadth. Feature extraction techniques discussed are principal component analysis, which transforms correlated features into linearly uncorrelated components capturing maximum variance. The document provides examples and steps for principal component analysis.
MACHINE LEARNING YEAR DL SECOND PART.pptxNAGARAJANS68
The document discusses various concepts related to machine learning models including prediction errors, overfitting, underfitting, bias, variance, hyperparameter tuning, and regularization techniques. It provides explanations of key terms and challenges in machine learning like the curse of dimensionality. Cross-validation methods like k-fold are presented as ways to evaluate model performance on unseen data. Optimization algorithms such as gradient descent and stochastic gradient descent are covered. Regularization techniques like Lasso, Ridge, and Elastic Net are introduced.
This document summarizes a proposed method for discriminative unsupervised dimensionality reduction called DUDR. It begins by introducing traditional dimensionality reduction techniques like PCA and LDA. It then discusses limitations of existing graph embedding methods that require constructing a graph beforehand. The proposed DUDR method jointly learns the graph construction and dimensionality reduction to avoid this dependency. It formulates an optimization problem to learn a projection matrix and affinity matrix simultaneously. Experimental results on synthetic and real-world datasets show DUDR achieves better clustering performance than other methods like PCA, LPP, k-means and NMF.
Principal Component Analysis (PCA) is an unsupervised learning algorithm used for dimensionality reduction. It transforms correlated variables into linearly uncorrelated variables called principal components. PCA works by considering the variance of each attribute to reduce dimensionality while preserving as much information as possible. It is commonly used for exploratory data analysis, predictive modeling, and visualization.
PCA is a dimensionality reduction technique that transforms a number of potentially correlated variables into a smaller number of uncorrelated variables called principal components. It works by geometrically projecting the data onto a lower dimensional space in a way that retains as much of the variation present in the original data as possible. PCA is widely used for applications like data visualization, feature extraction and dimensionality reduction. An example of using PCA in genetics is shown, where the variability in human DNA data can be strongly linked to geographical patterns through principal component analysis.
Feature extraction and selection are important techniques in machine learning. Feature extraction transforms raw data into meaningful features that better represent the data. This reduces dimensionality and complexity. Good features are unique to an object and prevalent across many data samples. Principal component analysis is an important dimensionality reduction technique that transforms correlated features into linearly uncorrelated principal components. This both reduces dimensionality and preserves information.
This document discusses the importance of data preprocessing techniques for improving data quality. It outlines several key steps in data preprocessing: data cleaning, which detects and corrects errors and inconsistencies; data integration, which merges data from multiple sources; data reduction, which reduces data dimensions or volumes; and data transformation, which converts data into appropriate forms for analysis. Specific techniques discussed include missing value imputation, binning, smoothing, normalization, and discretization. The overall goal of data preprocessing is to prepare raw data for data mining and ensure high quality results.
Machine learning lets you make better business decisions by uncovering patterns in your consumer behavior data that is hard for the human eye to spot. You can also use it to automate routine, expensive human tasks that were previously not doable by computers. In the business to business space (B2B), if your competitors can make wiser business decisions based on data and automate more business operations but you still base your decisions on guesswork and lack automation, you will lose out on business productivity. In this introduction to machine learning tech talk, you will learn how to use machine learning even if you do not have deep technical expertise on this technology.
Topics covered:
1.What is machine learning
2.What is a typical ML application architecture
3.How to start ML development with free resource links
4.Key decision factors in ML technology selection depending on use case scenarios
Feature engineering is the process of selecting, modifying, or creating new features (variables) from raw data to improve the performance of machine learning models. It involves identifying the most relevant features, transforming data into a suitable format, handling missing values, encoding categorical variables, scaling numerical data, and creating interaction terms or derived features. Effective feature engineering can significantly enhance a model's accuracy and interpretability by providing it with the most informative inputs. It is often considered a crucial step in the machine learning pipeline.
Introduction to Data Structures Sorting and searchingMvenkatarao
This document provides an overview of data structures and algorithms. It begins by defining a data structure as a way of storing and organizing data in a computer so that it can be used efficiently by algorithms. Data structures can be primitive, directly operated on by machine instructions, or non-primitive, developed from primitive structures. Linear structures maintain adjacency between elements while non-linear do not. Common operations on data structures include adding, deleting, traversing, sorting, searching, and updating elements. The document also defines algorithms and their properties, including finiteness, definiteness, inputs, outputs, and effectiveness. It discusses analyzing algorithms based on time and space complexity and provides examples of different complexities including constant, logarithmic, linear, quadratic,
Lecture - 10 Transformer Model, Motivation to Transformers, Principles, and ...Maninda Edirisooriya
Learn about the limitations of earlier Deep Sequence Models like RNNs, GRUs and LSTMs; Evolution of Attention Model as the Transformer Model with the paper, "Attention is All You Need". This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2024 first half of the year.
Learn End-to-End Learning, Multi-Task Learning, Transfer Learning and Meta Learning. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2024 first half of the year.
Learn Recurrent Neural Networks (RNN), GRU and LSTM networks and their architecture. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2024 first half of the year.
Extra Lecture - Support Vector Machines (SVM), a lecture in subject module St...Maninda Edirisooriya
Support Vector Machines are one of the main tool in classical Machine Learning toolbox. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Lecture 11 - KNN and Clustering, a lecture in subject module Statistical & Ma...Maninda Edirisooriya
Supervised ML technique, K-Nearest Neighbor and Unsupervised Clustering techniques are learnt in this lesson. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Lecture 10 - Model Testing and Evaluation, a lecture in subject module Statis...Maninda Edirisooriya
Model Testing and Evaluation is a lesson where you learn how to train different ML models with changes and evaluating them to select the best model out of them. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Lecture 9 - Decision Trees and Ensemble Methods, a lecture in subject module ...Maninda Edirisooriya
Decision Trees and Ensemble Methods is a different form of Machine Learning algorithm classes. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Lecture 6 - Logistic Regression, a lecture in subject module Statistical & Ma...Maninda Edirisooriya
Logistic Regression is the first non-linear classification ML algorithm taught in this course. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Lecture 5 - Gradient Descent, a lecture in subject module Statistical & Machi...Maninda Edirisooriya
Gradient Descent is the most commonly used learning algorithm for learning, including Deep Neural Networks with Back Propagation. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Lecture 4 - Linear Regression, a lecture in subject module Statistical & Mach...Maninda Edirisooriya
Simplest Machine Learning algorithm or one of the most fundamental Statistical Learning technique is Linear Regression. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Lecture 3 - Exploratory Data Analytics (EDA), a lecture in subject module Sta...Maninda Edirisooriya
Exploratory Data Analytics (EDA) is a data Pre-Processing, manual data summarization and visualization related discipline which is an earlier phase of data processing. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Lecture 2 - Introduction to Machine Learning, a lecture in subject module Sta...Maninda Edirisooriya
Introduction to Statistical and Machine Learning. Explains basics of ML, fundamental concepts of ML, Statistical Learning and Deep Learning. Recommends the learning sources and techniques of Machine Learning. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Analyzing the effectiveness of mobile and web channels using WSO2 BAMManinda Edirisooriya
This document summarizes a presentation about using WSO2 BAM to analyze the effectiveness of mobile and web channels for e-commerce. It discusses how both channels have advantages and are growing in popularity for business applications. WSO2 BAM is presented as a solution to monitor usage patterns of both channels, including user behavior, interactions, and preferences. A demo is shown of WSO2 BAM monitoring an online ticket booking system with both a web app and mobile app to analyze and compare usage of each channel.
WSO2 BAM is a big data analytics and monitoring tool that provides scalable data flow, storage, and processing. It allows users to publish data, analyze and summarize it using Hadoop clusters and Cassandra storage, and then visualize the results. The document discusses WSO2 BAM's architecture and configuration options. It also describes its out-of-the-box monitoring and analytics solutions for services, APIs, applications, and platforms.
The document provides an overview of the training organization Zone24x7. It describes Zone24x7 as a technological company that provides hardware and software solutions. It details Zone24x7's organizational structure, products and services, partners and clients, and an assessment of its current position including strengths, weaknesses and suggestions. The training experience involved working on various software development projects at Zone24x7 to gain exposure to tools, technologies and company practices.
The document is a final project report submitted by four students for their Bachelor's degree. It presents the Geo-Data Visualization Framework (GViz) developed as part of the project. The framework enables visualization of geospatial data on the web using existing JavaScript APIs and libraries. It describes the design and implementation of GViz over multiple iterations to address common challenges in visualizing geographic data.
This document provides an overview of a remotely operated toy car project. It outlines the main requirements, functionality, features, implementation challenges, production process, and marketing plan. The key requirements are for the car to be operated remotely via a computer and wireless camera. Functionality includes transmitting control signals from the computer to a receiver and microcontroller in the car. Challenges include minimizing circuit size and integrating components. The production process involves specialized team roles, programming, and interfacing. Marketing targets children and emphasizes the affordable price and attractive design.
This document describes an encryption Chrome extension for online chat. The extension encrypts chat text using 128-bit AES encryption with a common password between users. It was created to provide a cheap, private chat solution without third parties analyzing conversations or filtering keywords. The extension works by encrypting text on one end, sending the encrypted ciphertext over the network, and decrypting it on the other end. It sets passwords by hashing them to generate an encryption key and encrypts/decrypts text by breaking it into blocks and applying the AES cipher. The document demonstrates how to use the extension for encrypted chat and discusses its limitations, such as an inability to send emojis or a key sharing mechanism.
Forensic Science – Digital Forensics – Digital Evidence – The Digital Forensi...ManiMaran230751
Forensic Science – Digital Forensics – Digital Evidence – The Digital Forensics Process – Introduction – The
Identification Phase – The Collection Phase – The Examination Phase – The Analysis Phase – The
Presentation Phase.
Better Builder Magazine brings together premium product manufactures and leading builders to create better differentiated homes and buildings that use less energy, save water and reduce our impact on the environment. The magazine is published four times a year.
Optimize Indoor Air Quality with Our Latest HVAC Air Filter Equipment Catalogue
Discover our complete range of high-performance HVAC air filtration solutions in this comprehensive catalogue. Designed for industrial, commercial, and residential applications, our equipment ensures superior air quality, energy efficiency, and compliance with international standards.
📘 What You'll Find Inside:
Detailed product specifications
High-efficiency particulate and gas phase filters
Custom filtration solutions
Application-specific recommendations
Maintenance and installation guidelines
Whether you're an HVAC engineer, facilities manager, or procurement specialist, this catalogue provides everything you need to select the right air filtration system for your needs.
🛠️ Cleaner Air Starts Here — Explore Our Finalized Catalogue Now!
This presentation provides a detailed overview of air filter testing equipment, including its types, working principles, and industrial applications. Learn about key performance indicators such as filtration efficiency, pressure drop, and particulate holding capacity. The slides highlight standard testing methods (e.g., ISO 16890, EN 1822, ASHRAE 52.2), equipment configurations (such as aerosol generators, particle counters, and test ducts), and the role of automation and data logging in modern systems. Ideal for engineers, quality assurance professionals, and researchers involved in HVAC, automotive, cleanroom, or industrial filtration systems.
Tesia Dobrydnia brings her many talents to her career as a chemical engineer in the oil and gas industry. With the same enthusiasm she puts into her work, she engages in hobbies and activities including watching movies and television shows, reading, backpacking, and snowboarding. She is a Relief Senior Engineer for Chevron and has been employed by the company since 2007. Tesia is considered a leader in her industry and is known to for her grasp of relief design standards.
Design of a Hand Rehabilitation Device for Post-Stroke Patients..pptxyounisalsadah
Designing a hand rehabilitation device for post-stroke patients. Stimulation is achieved through movement and control via a program on a mobile phone. The fingers are not involved in the movement, as this is a separate project.
This presentation showcases a detailed catalogue of testing solutions aligned with ISO 4548-9, the international standard for evaluating the anti-drain valve performance in full-flow lubricating oil filters used in internal combustion engines.
Topics covered include:
Structural Health and Factors affecting.pptxgunjalsachin
Structural Health- Factors affecting Health of Structures,
Causes of deterioration in RC structures-Permeability of concrete, capillary porosity, air voids, Micro cracks and macro cracks, corrosion of reinforcing bars, sulphate attack, alkali silica reaction
Causes of deterioration in Steel Structures: corrosion, Uniform deterioration, pitting, crevice, galvanic, laminar, Erosion, cavitations, fretting, Exfoliation, Stress, causes of defects in connection
Maintenance and inspection of structures.
DIY Gesture Control ESP32 LiteWing Drone using PythonCircuitDigest
Build a gesture-controlled LiteWing drone using ESP32 and MPU6050. This presentation explains components, circuit diagram, assembly steps, and working process.
Read more : https://circuitdigest.com/microcontroller-projects/diy-gesture-controlled-drone-using-esp32-and-python-with-litewing
Ideal for DIY drone projects, robotics enthusiasts, and embedded systems learners. Explore how to create a low-cost, ESP32 drone with real-time wireless gesture control.
UNIT-1-PPT-Introduction about Power System Operation and ControlSridhar191373
Power scenario in Indian grid – National and Regional load dispatching centers –requirements of good power system - necessity of voltage and frequency regulation – real power vs frequency and reactive power vs voltage control loops - system load variation, load curves and basic concepts of load dispatching - load forecasting - Basics of speed governing mechanisms and modeling - speed load characteristics - regulation of two generators in parallel.
Module4: Ventilation
Definition, necessity of ventilation, functional requirements, various system & selection criteria.
Air conditioning: Purpose, classification, principles, various systems
Thermal Insulation: General concept, Principles, Materials, Methods, Computation of Heat loss & heat gain in Buildings
Lecture 8 - Feature Engineering and Optimization, a lecture in subject module Statistical & Machine Learning
1. DA 5230 – Statistical & Machine Learning
Lecture 8 – Feature Engineering and
Optimization
Maninda Edirisooriya
[email protected]
2. Features
• In general Features are X values/Independent Variables or Predictor
Variables of a Dataset
• Features can be
• Numerical values
• Categorical labels
• Complex structures like texts or images
• Having high quality (with more and relevant information) and independent
(with information not shared with other features) features can improve
model accuracy
• Having lower quality and highly correlated (less independent) features can
reduce model accuracy (due to noise) and increase computational burden
3. Feature Selection
• When a dataset is given, first all the non-
related features (columns) have to be
deleted as discussed in EDA
• Then you will find that you can increase
the number of related features arbitrarily
larger with feature engineering
• E.g.: Polynomial Regression feature
generation: Convert X1 and X2 into features
of X1, X2, X1X2, X1
2, X2
2
• Adding new features may reduce the
training set error but you will notice that
the test set error gets increased after a
certain level Source: https://stats.stackexchange.com/questions/184103/why-the-
error-on-a-training-set-is-decreasing-while-the-error-on-the-validation
4. Feature Selection
• Therefore, you have to find the optimum features that minimizes the test
set error
• This process is known as Feature Selection
• When there are n number of candidate features there are
𝒏!
𝒓! 𝒏−𝒓 !
different
ways of selecting r number of features
• As the optimum r can be any number, the search space for all possible r
becomes
𝒓=𝟏
𝒏
𝒏!
𝒓! 𝒏−𝒓 !
which grows exponentially with n
• This is known as the Curse of Dimensionality
• Forward Selection or Backward Elimination algorithm can be used to
select features without this exponential search space growth
5. Forward Selection
• In Forward Selection, we start with an empty set of features
• In each iteration we add the best feature to the model feature set so
that the model performance is increased in the test set
• Here the model performance increase in the test set is used as the evaluation
criteria of the algorithm
• If all the features are added OR if there is no new feature remaining
that increases the model performance when added, stop the
algorithm
• This is the stopping criteria of the algorithm
6. Backward Elimination
• In Backward Elimination we start with all the available features
• In each iteration we remove the worst feature from the model feature
set so that the model performance is increased in the test set
• Here the model performance increase in the test set is used as the evaluation
criteria of the algorithm
• If all the features are removed OR if there is no existing feature
remaining that increases the model performance when removed,
stop the algorithm
• This is the stopping criteria of the algorithm
7. Common Nature of these Algorithms
• These algorithms are faster than the pure Feature Selection
• In these algorithms, the evaluation criteria and stopping criteria can
be customized as you like
• E.g.: Maximum/minimum number of features can also be used as the
stopping criteria as well
• Cross-validation performance increase can be used as the evaluation criteria
when the dataset is small
• Because these are heuristic algorithms, we may miss some better
feature combinations which may result better performance
• That is what we sacrifice for the speed of these new algorithms
8. Feature Transformation
Numerical features may exist with unwanted distributions
For example, some X values in a dataset for a Linear Regression can be non-
linear which can be transformed to a linear relationship using a higher degree
of that variable
X1
Y Y
X1
2
Transformation
X1 → X1
2 or
X1 → exp(X1)
9. Feature Transformation
Non-normal frequency distributions can be converted to normal
distributions like follows
X1
X1
2
n’th root OR log(X1)
Right skewed
Distribution
Normal
Distribution
X1
Frequency
X1
2
n’th power OR exp(X1)
Left skewed
Distribution
Normal
Distribution
Frequency
Frequency
Frequency
10. Feature Encoding
• Many machine learning algorithms need numerical values for their X
variables
• Therefore categorical variables have to be converted into numerical
variables, to be used them as model features
• There are many ways to encode categorical variables to numerical
• Nominal variables (e.g.: Color, Gender) are generally encoded with
One-Hot Encoding
• Ordinal variables (e.g.: T-shirt size, Age group) are generally encoded
with Label Encoding
13. Scaling Features
• Numerical data features of a dataset can have different scales
• E.g.: Number of bedrooms in a house may spread between 1 to 5 while the
square feet of a house can spread between 500 to 4000
• When these features are used as they are, there can be problems
when taking vector distances between each other
• E.g.: Can affect the convergence rate in Gradient Descent algorithm
• When regularization is applied, most L1 and L2 regularization
components are applied in the same scale for all the features
• i.e.: Small scale features are highly regularized and vice versa
• Interpreting a model can be difficult, as model parameter scales can
be affected by the feature’s scale
14. Scaling Features
• Therefore, it is better all the numerical features of the model to be
scaled to a single scale
• E.g.: 0 to 1 scale
• There are 2 main widely used forms of scales
1. Normalization
2. Standardization
15. Normalization
• In Normalization all features are transformed to a feature with a fixed
range from 0 to 1
• Every feature is scaled taking the difference between the maximum
and the minimum X values of the feature as 1
• Each data point, Xi can be scaled as, (where min(X) is the minimum X
value and max(X) is the maximum X value of that feature)
Xi =
𝐗𝐢−𝐦𝐢𝐧(𝐗)
𝐦𝐚𝐱 𝐗 −𝐦𝐢𝐧(𝐗)
16. Standardization
• In Standardization all the features are transformed to a standard
normal distribution
• Every feature is scaled assuming the distribution is normal
• Here ഥ
𝐗 is the mean and 𝝈 is the standard deviation of the feature
• Each data point of the feature, Xi can be scaled as,
Xi =
𝐗𝐢−ഥ
𝐗
𝝈
17. Handling Missing Data Values in Features
• In a practical dataset there can be values missing in some data fields
due to different reasons
• Most Machine Learning algorithms cannot handle empty or nil valued
data values
• Therefore, the missing values have to be either
• Removed along with its data row OR with its data column OR
• Filled with an approximate value which is also known as Imputation
18. Filling a Missing Value (Imputation)
• A missing value actually represents the unavailability of information
• But we can fill them with a predicted value approximating its original
value (i.e. Imputation)
• Remember that filling a missing value does not introduce any new
information to the dataset unless it is predicted by another intelligent
system
• Therefore, if possible, if the number of missing values are significantly
high in a certain data row or a column, it may be better to remove the
whole data row or the column
19. Imputation Techniques
• Mean/Median/Mode Imputation
• The missing value can be replaced with the best Central Tendency measure
best suitable for the feature data distribution
• If the distribution is Normal, Mean can be used for imputation
• If the distribution is not Normal, Median can be used
• Forward/Backward Fill
• Filling the missing value with the previous known value of the same column in
a timeseries or ordered dataset is known as the Forward Fill
• Filling the missing value with the next known value of the same column in a
timeseries or ordered dataset is known as the Backward Fill
• Interpolation can be used to predict the missing value using the
known previous and subsequent values
20. Imputation Techniques
• Machine Learning techniques can also be used to predict the missing
value
• E.g.: Linear Regression, K-Nearest Neighbor algorithm
• When the probability distribution is known, a random number from
the distribution can be generated to fill the missing value as well
• In some cases missing values may follow a different distribution from
the available data distribution
• E.g.: When medical data is collected from a form, missing values of being a
smoker (binary value) may be biased towards being a smoker
21. Feature Generation
• Generating new features using the existing features is a way of
making the useful information available to the model
• As the existing features are used to generate new features, no new
information is really introduced to the Machine Learning model, but
new features may uncover hidden information in the dataset to the
ML model
• Domain knowledge about the problem to be solved using ML, is
important at the Feature Generation
22. Feature Generation Techniques
• Polynomial Features
• Involves creating new features by changing the power of an existing feature
• E.g.: X1 → X1
2, X1
3
• Interaction Features
• Combines several features to create a new feature
• E.g.: Multiply length and width of a land in a model where the land price is to be
predicted
• Binning Features
• Groups numerical features into bins or intervals
• E.g.: Convert age parameter into age-groups
• Converts numerical variables into categorical variables
• Helps to reduce the noise and overfitting
23. One Hour Homework
• Officially we have one more hour to do after the end of the lecture
• Therefore, for this week’s extra hour you have a homework
• Feature Engineering is a Data Science related subject to be mastered by
anyone who is interested in ML, which can help to improve the accuracy of a
ML model significantly!
• There are many more Feature Engineering techniques and it is very useful to
learn them and understanding why they are used with clear reasons
• Once you have completed studying the given set of techniques, search about
other techniques as well
• Good Luck!