Concept learning and candidate elimination algorithmswapnac12
This document discusses concept learning, which involves inferring a Boolean-valued function from training examples of its input and output. It describes a concept learning task where each hypothesis is a vector of six constraints specifying values for six attributes. The most general and most specific hypotheses are provided. It also discusses the FIND-S algorithm for finding a maximally specific hypothesis consistent with positive examples, and its limitations in dealing with noise or multiple consistent hypotheses. Finally, it introduces the candidate-elimination algorithm and version spaces as an improvement over FIND-S that can represent all consistent hypotheses.
Concept Learning presentation with benefits and usesmousmiin
Concept learning is a process of acquiring knowledge or understanding about specific concepts or categories from examples. In essence, it's the ability of a machine learning model to identify patterns and generalize from data in order to classify new instances correctly.
Concept learning is a process of acquiring knowledge or understanding about specific concepts or categories from examples. In essence, it's the ability of a machine learning model to identify patterns and generalize from data in order to classify new instances correctly.
1. The document discusses machine learning concepts including what learning is, how to construct programs that automatically improve with experience, and designing learning systems.
2. It provides examples of learning problems involving chess games and handwriting recognition to classify days that a friend enjoys water sports.
3. Concept learning algorithms like FIND-S and version space algorithms like Candidate Elimination are introduced to learn concepts from examples in a restricted hypothesis space.
1) The document discusses concept learning, which involves inferring a Boolean function from training examples. It focuses on a concept learning task where hypotheses are represented as vectors of constraints on attribute values.
2) It describes the FIND-S algorithm, which finds the most specific hypothesis consistent with positive examples by generalizing constraints. However, FIND-S has limitations like ignoring negative examples.
3) The Candidate-Elimination algorithm represents the version space of all hypotheses consistent with examples to address FIND-S limitations. It outputs the version space rather than a single hypothesis.
concept-learning of artificial intelligencessuser01fa1b
This document discusses concept learning through inductive logic. It introduces the concepts of:
- Training examples that provide positive and negative examples of a target concept
- Hypotheses that attempt to represent the target concept in the form of logical rules
- The version space that contains all hypotheses consistent with the training examples
- Candidate elimination, an algorithm that iteratively generalizes and specializes hypotheses within the version space based on new training examples
- The need for an appropriate hypothesis space and inductive bias to allow learning from a limited number of examples and enable generalization to new cases.
The document discusses concept learning through inductive logic. It introduces the concept learning task of predicting when a person will enjoy a sport based on attributes of the day. It describes representing hypotheses as conjunctions of attribute values and the version space approach of tracking the most specific and most general consistent hypotheses. The document explains the candidate elimination algorithm, which uses positive and negative examples to generalize the specific boundary and specialize the general boundary, respectively, until the version space is fully resolved.
The document discusses different approaches for concept learning from examples, including viewing it as a search problem to find the hypothesis that best fits the training examples. It also describes the general-to-specific learning approach, where the goal is to find the maximally specific hypothesis consistent with the positive training examples by starting with the most general hypothesis and replacing constraints to better fit the examples. The document also discusses the version space and candidate elimination algorithms for obtaining the version space of all hypotheses consistent with the training data.
The document discusses concept learning and the candidate elimination algorithm. It defines concept learning as inducing a function that maps examples into categories. It then describes concept learning as a search problem to find the most specific hypothesis consistent with training examples. The candidate elimination algorithm maintains the most specific and general hypotheses consistent with examples to represent the version space of possible concepts. It updates these lists by generalizing specific hypotheses or specializing general hypotheses based on new positive and negative examples.
This document discusses machine learning concepts of concept learning and decision-tree learning. It describes concept learning as inferring a boolean function from training examples and using algorithms like Candidate Elimination to search the hypothesis space. Decision tree learning is explained as representing classification functions as trees with nodes testing attributes, allowing disjunctive concepts. The ID3 algorithm is presented as a greedy top-down search that selects the best attribute at each node using information gain, potentially overfitting data without pruning or a validation set.
The document discusses concepts and concept learning in machine learning. It begins by defining concepts as functions that map inputs to outputs. The goal of concept learning is to infer this underlying concept function from examples. It introduces the key components of concept learning:
1. The instance space containing labeled examples
2. The concept space containing all possible concept functions
3. The hypothesis space which reduces the concept space by introducing biases or constraints
4. Training examples which are used to search the hypothesis space for the best fitting concept
It then provides examples to illustrate these components using a medical domain about predicting whether a patient is sick based on attributes like temperature and blood pressure. It describes how the hypothesis space can be ordered from general to
The document summarizes key concepts in machine learning including concept learning as search, general-to-specific learning, version spaces, candidate elimination algorithm, and decision trees. It discusses how concept learning can be viewed as searching a hypothesis space to find the hypothesis that best fits the training examples. The candidate elimination algorithm represents the version space using the most general and specific hypotheses to efficiently learn from examples.
This document provides an overview and syllabus for CS 446: Machine Learning, taught by Gerald DeJong. Key details include:
- The course will use Mitchell's Machine Learning textbook
- Important dates: midterm on Oct 4, final on Dec 12
- Homework and projects will be submitted in class, with late penalties of 20% per day up to 3 days late
- Topics covered will include decision trees, linear threshold units, probabilistic representations, reinforcement learning, clustering, and more based on student interest
This document provides an overview and syllabus for CS 446: Machine Learning, taught by Gerald DeJong. Key details include:
- The course will use Mitchell's Machine Learning textbook
- Important dates: midterm on Oct 4, final on Dec 12
- Homeworks and projects will be submitted in class, with late penalties of 20% per day up to 3 days late
- Topics covered will include decision trees, linear threshold units, probabilistic representations, reinforcement learning, clustering, and more based on student interest
The document discusses concept learning algorithms. It introduces the problem of concept learning as inducing a function to classify examples into categories based on their attributes. The Candidate Elimination Algorithm (CEA) is presented as a method for finding all hypotheses consistent with training examples without enumerating them. CEA works by maintaining the most specific (S) and most general (G) consistent hypotheses. It updates S and G in response to positive and negative examples.
The document discusses version space learning, an approach to machine learning where both the most general and most specific hypotheses consistent with the training examples are maintained. It begins by introducing concept learning and version spaces, showing how all possible hypotheses can be represented as a lattice. The Find-S and Dual Find-S algorithms are presented for updating the version spaces in response to positive and negative examples. The key properties of version spaces are that they track all hypotheses consistent with the examples seen so far, avoiding premature commitment to a single hypothesis.
The document discusses different approaches for concept learning from examples, including viewing it as a search problem to find the hypothesis that best fits the training examples. It also describes the general-to-specific learning approach, where the goal is to find the maximally specific hypothesis consistent with the positive training examples by starting with the most general hypothesis and replacing constraints to better fit the examples. The document also discusses the version space and candidate elimination algorithms for obtaining the version space of all hypotheses consistent with the training data.
The document discusses concept learning and the candidate elimination algorithm. It defines concept learning as inducing a function that maps examples into categories. It then describes concept learning as a search problem to find the most specific hypothesis consistent with training examples. The candidate elimination algorithm maintains the most specific and general hypotheses consistent with examples to represent the version space of possible concepts. It updates these lists by generalizing specific hypotheses or specializing general hypotheses based on new positive and negative examples.
This document discusses machine learning concepts of concept learning and decision-tree learning. It describes concept learning as inferring a boolean function from training examples and using algorithms like Candidate Elimination to search the hypothesis space. Decision tree learning is explained as representing classification functions as trees with nodes testing attributes, allowing disjunctive concepts. The ID3 algorithm is presented as a greedy top-down search that selects the best attribute at each node using information gain, potentially overfitting data without pruning or a validation set.
The document discusses concepts and concept learning in machine learning. It begins by defining concepts as functions that map inputs to outputs. The goal of concept learning is to infer this underlying concept function from examples. It introduces the key components of concept learning:
1. The instance space containing labeled examples
2. The concept space containing all possible concept functions
3. The hypothesis space which reduces the concept space by introducing biases or constraints
4. Training examples which are used to search the hypothesis space for the best fitting concept
It then provides examples to illustrate these components using a medical domain about predicting whether a patient is sick based on attributes like temperature and blood pressure. It describes how the hypothesis space can be ordered from general to
The document summarizes key concepts in machine learning including concept learning as search, general-to-specific learning, version spaces, candidate elimination algorithm, and decision trees. It discusses how concept learning can be viewed as searching a hypothesis space to find the hypothesis that best fits the training examples. The candidate elimination algorithm represents the version space using the most general and specific hypotheses to efficiently learn from examples.
This document provides an overview and syllabus for CS 446: Machine Learning, taught by Gerald DeJong. Key details include:
- The course will use Mitchell's Machine Learning textbook
- Important dates: midterm on Oct 4, final on Dec 12
- Homework and projects will be submitted in class, with late penalties of 20% per day up to 3 days late
- Topics covered will include decision trees, linear threshold units, probabilistic representations, reinforcement learning, clustering, and more based on student interest
This document provides an overview and syllabus for CS 446: Machine Learning, taught by Gerald DeJong. Key details include:
- The course will use Mitchell's Machine Learning textbook
- Important dates: midterm on Oct 4, final on Dec 12
- Homeworks and projects will be submitted in class, with late penalties of 20% per day up to 3 days late
- Topics covered will include decision trees, linear threshold units, probabilistic representations, reinforcement learning, clustering, and more based on student interest
The document discusses concept learning algorithms. It introduces the problem of concept learning as inducing a function to classify examples into categories based on their attributes. The Candidate Elimination Algorithm (CEA) is presented as a method for finding all hypotheses consistent with training examples without enumerating them. CEA works by maintaining the most specific (S) and most general (G) consistent hypotheses. It updates S and G in response to positive and negative examples.
The document discusses version space learning, an approach to machine learning where both the most general and most specific hypotheses consistent with the training examples are maintained. It begins by introducing concept learning and version spaces, showing how all possible hypotheses can be represented as a lattice. The Find-S and Dual Find-S algorithms are presented for updating the version spaces in response to positive and negative examples. The key properties of version spaces are that they track all hypotheses consistent with the examples seen so far, avoiding premature commitment to a single hypothesis.
NumPy is a Python library that provides multidimensional array and matrix objects to perform scientific computing. It contains efficient functions for operations on arrays like arithmetic, aggregation, copying, indexing, slicing, and reshaping. NumPy arrays have advantages over native Python sequences like fixed size and efficient mathematical operations. Common NumPy operations include elementwise arithmetic, aggregation functions, copying and transposing arrays, changing array shapes, and indexing/slicing arrays.
This document discusses SQL queries on sample student and department databases. It covers basic queries using SELECT, FROM, and WHERE clauses. It also demonstrates functions like LIKE, UNION, JOIN, ORDER BY, and set operations. Examples retrieve names of students in a certain semester or department, union results from multiple tables, perform left joins, and order query results.
Discrete mathematics is the study of mathematical structures that are discrete rather than continuous. It has many applications in computer science, as computers use discrete structures to represent and manipulate data. Some key topics in discrete mathematics include logic and proofs, number theory, counting, and graph theory. These topics help with problem solving, algorithms, data structures, cryptography, and the design of efficient computer systems and networks. Examples of problems solved with discrete mathematics include the fastest sorting algorithms, map coloring, efficient data transmission, electronic health records, and Google Maps routing.
The document discusses machine language instruction formats. It covers:
1) Instruction formats have optional prefixes that modify register and address sizes, and opcodes that select operations.
2) The MOD-REG-R/M field specifies addressing modes like register, direct memory with/without displacement.
3) Segment overrides add a prefix code before instructions. 64-bit modes add REX prefixes to address more registers.
4) MOV instructions are discussed, including immediate to register/memory formats and their byte codes.
The document discusses direct memory access (DMA) and interrupts. It describes how DMA allows direct data transfer between memory and I/O devices without involving the CPU. This is handled by a DMA controller through a request-grant handshake using HOLD and HLDA pins. The document also categorizes different types of interrupts like hardware, software and exceptions. It explains how interrupts alter program flow and are serviced by interrupt service routines (ISRs) through an interrupt vector table.
The 8255 Programmable Peripheral Interface chip features 3 8-bit I/O ports that can be programmed to operate in different modes. It has input and output pins to interface with a microprocessor. The chip can be programmed to set I/O modes for each port independently using control words that define the port configurations and functions.
System modeling and simulation involves creating simplified representations of real-world systems to understand and evaluate their behavior over time. A system is composed of interconnected parts designed to achieve specific objectives. A model abstracts and simplifies a system for analysis. Simulation executes a model over time to observe how a system operates. It allows experimenting with systems that may be too expensive, dangerous or complex to study directly. Simulation has many uses including analyzing systems before implementation, optimizing designs, training, and evaluating "what-if" scenarios. Key areas where simulation is applied include manufacturing, business, healthcare, transportation and the military.
The document discusses a proposed settlement agreement between two parties, John Doe and Richard Roe, to resolve a legal dispute over an automobile accident. The agreement states that John Doe will pay Richard Roe $5,000 in damages and that both parties will dismiss all claims against each other to avoid further legal proceedings. In exchange for the payment and dismissal of claims, Richard Roe agrees to release John Doe from any liability related to the accident.
This document is a scanned receipt from a grocery store purchase on January 15th, 2022 for $58.64. It lists the items bought which include produce, dairy, baked goods, and other grocery items. The payment was made with a credit card ending in 4321.
Peak ground acceleration (PGA) is a critical parameter in ground-motion investigations, in particular in earthquake-prone areas such as Iran. In the current study, a new method based on particle swarm optimization (PSO) is developed to obtain an efficient attenuation relationship for the vertical PGA component within the northern Iranian plateau. The main purpose of this study is to propose suitable attenuation relationships for calculating the PGA for the Alborz, Tabriz and Kopet Dag faults in the vertical direction. To this aim, the available catalogs of the study area are investigated, and finally about 240 earthquake records (with a moment magnitude of 4.1 to 6.4) are chosen to develop the model. Afterward, the PSO algorithm is used to estimate model parameters, i.e., unknown coefficients of the model (attenuation relationship). Different statistical criteria showed the acceptable performance of the proposed relationships in the estimation of vertical PGA components in comparison to the previously developed relationships for the northern plateau of Iran. Developed attenuation relationships in the current study are independent of shear wave velocity. This issue is the advantage of proposed relationships for utilizing in the situations where there are not sufficient shear wave velocity data.
Video Games and Artificial-Realities.pptxHadiBadri1
🕹️ #GameDevs, #AIteams, #DesignStudios — I’d love for you to check it out.
This is where play meets precision. Let’s break the fourth wall of slides, together.
As an AI intern at Edunet Foundation, I developed and worked on a predictive model for weather forecasting. The project involved designing and implementing machine learning algorithms to analyze meteorological data and generate accurate predictions. My role encompassed data preprocessing, model selection, and performance evaluation to ensure optimal forecasting accuracy.
Module4: Ventilation
Definition, necessity of ventilation, functional requirements, various system & selection criteria.
Air conditioning: Purpose, classification, principles, various systems
Thermal Insulation: General concept, Principles, Materials, Methods, Computation of Heat loss & heat gain in Buildings
Expansive soils (ES) have a long history of being difficult to work with in geotechnical engineering. Numerous studies have examined how bagasse ash (BA) and lime affect the unconfined compressive strength (UCS) of ES. Due to the complexities of this composite material, determining the UCS of stabilized ES using traditional methods such as empirical approaches and experimental methods is challenging. The use of artificial neural networks (ANN) for forecasting the UCS of stabilized soil has, however, been the subject of a few studies. This paper presents the results of using rigorous modelling techniques like ANN and multi-variable regression model (MVR) to examine the UCS of BA and a blend of BA-lime (BA + lime) stabilized ES. Laboratory tests were conducted for all dosages of BA and BA-lime admixed ES. 79 samples of data were gathered with various combinations of the experimental variables prepared and used in the construction of ANN and MVR models. The input variables for two models are seven parameters: BA percentage, lime percentage, liquid limit (LL), plastic limit (PL), shrinkage limit (SL), maximum dry density (MDD), and optimum moisture content (OMC), with the output variable being 28-day UCS. The ANN model prediction performance was compared to that of the MVR model. The models were evaluated and contrasted on the training dataset (70% data) and the testing dataset (30% residual data) using the coefficient of determination (R2), Mean Absolute Error (MAE), and Root Mean Square Error (RMSE) criteria. The findings indicate that the ANN model can predict the UCS of stabilized ES with high accuracy. The relevance of various input factors was estimated via sensitivity analysis utilizing various methodologies. For both the training and testing data sets, the proposed model has an elevated R2 of 0.9999. It has a minimal MAE and RMSE value of 0.0042 and 0.0217 for training data and 0.0038 and 0.0104 for testing data. As a result, the generated model excels the MVR model in terms of UCS prediction.
DIY Gesture Control ESP32 LiteWing Drone using PythonCircuitDigest
Build a gesture-controlled LiteWing drone using ESP32 and MPU6050. This presentation explains components, circuit diagram, assembly steps, and working process.
Read more : https://circuitdigest.com/microcontroller-projects/diy-gesture-controlled-drone-using-esp32-and-python-with-litewing
Ideal for DIY drone projects, robotics enthusiasts, and embedded systems learners. Explore how to create a low-cost, ESP32 drone with real-time wireless gesture control.
Optimize Indoor Air Quality with Our Latest HVAC Air Filter Equipment Catalogue
Discover our complete range of high-performance HVAC air filtration solutions in this comprehensive catalogue. Designed for industrial, commercial, and residential applications, our equipment ensures superior air quality, energy efficiency, and compliance with international standards.
📘 What You'll Find Inside:
Detailed product specifications
High-efficiency particulate and gas phase filters
Custom filtration solutions
Application-specific recommendations
Maintenance and installation guidelines
Whether you're an HVAC engineer, facilities manager, or procurement specialist, this catalogue provides everything you need to select the right air filtration system for your needs.
🛠️ Cleaner Air Starts Here — Explore Our Finalized Catalogue Now!
world subdivision.pdf...................bmmederos12
Concept Learning - Find S Algorithm,Candidate Elimination Algorithm
1. Basics of Learning Theory
By :
Sharmila Chidaravalli
Assistant Professor
Global Academy of Technology
2. Learning is a process by which one can acquire knowledge and construct
new ideas or concepts based on the experiences.
Machine learning is an intelligent way of learning general concept from
training examples without writing a program.
There are many machine learning algorithms through which computers can
intelligently learn from past data or experiences, identify patterns, and
make predictions when new data is fed.
3. What is Concept Learning…?
Concept learning can be formulated as a problem of searching through a predefined
space of potential hypotheses for the hypothesis that best fits the training examples.
“A task of acquiring potential hypothesis (solution) that best fits the
given training examples.”
Concept learning requires three things:
1. Input
2. Output
3. Test
Formally, Concept learning is defined as–
"Given a set of hypotheses, the learner searches through the
hypothesis space to identify the best hypothesis that matches the
target concept".
4. Sky AirTemp Humidity Wind Water Forecast EnjoySport
Sunny Warm Normal Strong Warm Same Yes
Sunny Warm High Strong Warm Same Yes
Rainy Cold High Strong Warm Change No
Sunny Warm High Strong Cool Change Yes
Consider the example task of learning the target concept “days on which John enjoys his favorite
water sport.”
Below Table describes a set of example days, each represented by a set of attributes.
The attribute EnjoySport indicates whether or not John enjoys his favorite water sport on this day.
The task is to learn to predict the value of EnjoySport for an arbitrary day, based on the values of its
other attributes.
Objective is to learn
{ Sky AirTemp Humidity Wind Water Forecast } → EnjoySport
6. Concept Learning Task Notation
Sky AirTemp Humidity Wind Water Forecast EnjoySport
Sunny Warm Normal Strong Warm Same Yes
Sunny Warm High Strong Warm Same Yes
Rainy Cold High Strong Warm Change No
Sunny Warm High Strong Cool Change Yes
Training Examples (D)
Target Concept (C)
Training Instances (X)
Independent Variables Dependent Variables
7. Representation of a Hypothesis
A hypothesis ‘h’ approximates a target function ‘f ’ to represent the relationship
between the independent attributes and the dependent attribute of the training
instances.
The hypothesis is the predicted approximate model that best maps the inputs to
outputs.
Each hypothesis is represented as a conjunction of attribute conditions in the
antecedent part.
For example, (Tail = Short) ʌ (Color = Black)….
The set of hypothesis in the search space is called as hypotheses.
Hypotheses are the plural form of hypothesis.
Generally ‘H’ is used to represent the hypotheses and ‘h’ is used to represent a
candidate hypothesis.
8. What hypothesis representation is provided to the learner?
• Let’s consider a simple representation in which each hypothesis consists of a
conjunction of constraints on the instance attributes.
• Let each hypothesis be a vector of six constraints, specifying the values of the six
attributes Sky, AirTemp, Humidity, Wind, Water, and Forecast.
Representation of a Hypothesis
For each attribute, the hypothesis will either
•indicate by a “?’ that any value is acceptable for this attribute,
•specify a single required value (e.g., Warm) for the attribute, or
•indicate by a “ø” that no value is acceptable.
9. Representation of a Hypothesis
Most General and Specific Hypothesis
The most general hypothesis-that every day is a positive example-is represented by
(?, ?, ?, ?, ?, ?)
the most specific possible hypothesis-that no day is a positive example-is represented by
(ø, ø, ø, ø, ø, ø)
If some instance x satisfies all the constraints of hypothesis h, then h classifies x as a positive example (h(x) = 1).
To illustrate, the hypothesis that Person enjoys his favorite sport only on cold days with high humidity
(independent of the values of the other attributes) is represented by the expression
(?, Cold, High, ?, ?, ?)
10. • The set of items over which the concept is defined is called the set of instances, which is
denoted by X.
Example: X is the set of all possible days, each represented by the attributes:
Sky, AirTemp, Humidity, Wind, Water, and Forecast
• The concept or function to be learned is called the target concept, which is denoted by c.
c can be any Boolean valued function defined over the instances X c: X→ {0, 1}
Example: The target concept corresponds to the value of the attribute EnjoySport
(i.e., c(x) = 1 if EnjoySport = Yes, and c(x) = 0 if EnjoySport = No).
Notation
11. • Instances for which c(x) = 1 are called positive examples, or members of the target concept.
• Instances for which c(x) = 0 are called negative examples, or non-members of the target concept.
• The ordered pair (x, c(x)) to describe the training example consisting of the instance x and its
target concept value c(x).
• D to denote the set of available training examples
• The symbol H to denote the set of all possible hypotheses that the learner may consider regarding
the identity of the target concept.
Each hypothesis h in H represents a Boolean valued function defined over X h: X→{0, 1}
The goal of the learner is to find a hypothesis h such that h(x) = c(x) for all x in X.
Notation
12. • Given:
Instances X: Possible days, each described by the attributes
• Sky (with possible values Sunny, Cloudy, and Rainy),
• AirTemp (with values Warm and Cold),
• Humidity (with values Normal and High),
• Wind (with values Strong and Weak),
• Water (with values Warm and Cool),
• Forecast (with values Same and Change).
• Hypotheses H: Each hypothesis is described by a conjunction of constraints on the attributes Sky, AirTemp,
Humidity, Wind, Water, and Forecast. The constraints may be "?" (any value is acceptable), “Φ” (no value is
acceptable), or a specific value.
• Target concept c: EnjoySport : X → {0, l}
• Training examples D: Positive and negative examples of the target function
• Determine:
• A hypothesis h in H such that h(x) = c(x) for all x in X.
13. A CONCEPT LEARNING TASK
Concept learning can be viewed as the task of searching through
a large space of hypotheses implicitly defined by the hypothesis
representation.
The goal of this search is to find the hypothesis that best fits the training
examples.
It is important to note that by selecting a hypothesis representation, the designer of the learning
algorithm implicitly defines the space of all hypotheses that the program can ever represent and
therefore can ever learn.
14. Instance Space
Consider, for example, the instances X and hypotheses H in the EnjoySport learning task.
Given that the attribute Sky has three possible values, and that AirTemp, Humidity, Wind, Water,
and Forecast each have two possible values, the instance space X contains exactly 3 . 2 . 2 . 2 . 2 . 2 =
96 distinct instances.
15. Example:
Let’s assume there are two features F1 and F2 with F1 has A and B as possibilities and F2 as X
and Y as possibilities.
F1 – > A, B
F2 – > X, Y
Instance Space: (A, X), (A, Y), (B, X), (B, Y) – 4 Examples
Hypothesis Space: (A, X), (A, Y), (A, ø), (A, ?), (B, X), (B, Y), (B, ø), (B, ?), (ø, X), (ø, Y), (ø, ø), (ø,
?), (?, X), (?, Y), (?, ø), (?, ?) – 16
Hypothesis Space: (A, X), (A, Y), (A, ?), (B, X), (B, Y), (B, ?), (?, X), (?, Y (?, ?) – 10
16. Hypothesis Space
Hypothesis space is the set of all possible hypotheses that approximates the target function f.
Similarly there are 5 . 4 . 4 . 4 . 4 . 4 = 5120 syntactically distinct hypotheses within H.
Notice, however, that every hypothesis containing one or more “ø” symbols represents the empty set of
instances; that is, it classifies every instance as negative.
Therefore, the number of semantically distinct hypotheses is only 1 + (4 . 3 . 3 . 3 . 3 . 3) = 973.
EnjoySport example is a very simple learning task, with a relatively small, finite hypothesis space.
17. Find-S Algorithm
* Concept Learning
* Finds most specific hypothesis
* Considers only positive examples
General Hypothesis Specific Hypothesis
G={ ‘?’ , ’?’ , ‘?’ , ‘?’ , ‘?’ , ‘?’ } S = { ‘Ø ’ , ’ Ø ’ , ‘Ø ’ , ‘Ø ’ , ‘Ø ’ , ‘Ø ’ }
Attributes
Find Maximally Specific Hypothesis
18. Algorithm :
Step 1 : Initialize the hypothesis (h) to most specific hypothesis
Step 2 : for each +ve example:
for each attribute in the example:
if attribute value = hypothesis value :
Do Nothing
else
Replace its hypothesis value with the more general constraint ‘?’
19. Dataset
6 attributes (Nominal-valued (symbolic) attributes):
Sky (SUNNY, RAİNY, CLOUDY),
Temp (WARM,COLD),
Humidity (NORMAL, HIGH),
Wind (STRONG, WEAK),
Water (WARM, COOL),
Forecast (SAME, CHANGE)
Sky AirTemp Humidity Wind Water Forecast EnjoySport
Sunny Warm Normal Strong Warm Same Yes
Sunny Warm High Strong Warm Same Yes
Rainy Cold High Strong Warm Change No
Sunny Warm High Strong Cool Change Yes
31. Design of a Learning System
A system that is built around a learning algorithm is called a learning system. The design of systems focuses
on these steps:
1. Choosing a training experience
2. Choosing a target function
3. Representation of a target function
4. Function approximation Training Experience Let us consider designing of a chess
Choosing the Training Experience
• The first design choice is to choose the type of training experience from which the
system will learn.
• The type of training experience available can have a significant impact on success or
failure of the learner.
32. Determine the Target Function
The next step is the determination of a target function.
In this step, the type of knowledge that needs to be learnt is determined.
In direct experience, a board move is selected and is determined whether it is a good move or not
against all other moves.
If it is the best move, then it is chosen as: B -> M, where, B and M are legal moves.
In indirect experience, all legal moves are accepted and a score is generated for each.
The move with largest score is then chosen and executed.
33. Determine the Target Function Representation
The representation of knowledge may be a table, collection of rules or a neural network. The linear
combination of these factors can be coined as:
where, x1, x2 and x3 represent different board features and w0, w1, w2 and w3 represent weights.
34. Choosing an Approximation Algorithm for the Target Function
The focus is to choose weights and fit the given training samples effectively. The
aim is to reduce the error given as:
35. Problems with Find-S
Depending on H, there might be several maximally consistent hypotheses, and
there is no way for Find-S to find them.
All of them are equally likely.
There are several problems with the Find-S algorithm:
It cannot determine if it has learnt the concept.
There might be several other hypotheses that match as well – has
it found the only one?
It cannot detect when training data is inconsistent.
We would like to detect and be tolerant to errors and noise.
Why do we want the most specific hypothesis?
Some other hypothesis might be more useful.
36. Version Space
Definition:
A hypothesis h is consistent with a set of training examples D if and only if h(x)=c(x) for each example
<x , c ( x )> in D.
Version space = the subset of all hypotheses in H consistent with the training examples D.
Definition:
The version space, denoted VSH,D, with respect to hypothesis space H and training examples D,
is the subset of hypotheses from H consistent with the training examples in D.
37. Representation of Version Space
Option 1: List all of the members in the version space.
Works only when the hypothesis space H is finite!
Option 2: Store only the set of most general members G and the set of most specific
members S.
Given these two sets, it is possible to generate any member of the
version space as needed.
38. List-Eliminate Algorithm
2. For each training example, <x , c ( x )>
remove from VSH,D any hypothesis that is
inconsistent with the training example h(x) ≠ c(x)
1. VSH,D a list containing every hypothesis in H
3. Output the list of hypotheses in VSH,D
Advantage: Guaranteed to output all hypotheses
consistent with the training examples.
But inefficient! Even in this simple example, there are
1+4·3·3·3·3 = 973 semantically distinct hypotheses.
39. Candidate-Elimination Algorithm
G maximally general hypothesis in H
S maximally specific hypothesis in H
For each training example d = <x , c ( x )>
modify G and S so that G and S are consistent with d
42. Candidate-Elimination Algorithm (detailed)
G maximally general hypothesis in H
S maximally specific hypothesis in H
For each training example d = <x , c ( x )>
If d is a positive example
• Remove from G any hypothesis that is inconsistent with d
• For each hypothesis s in S that is not consistent with d
• Remove s from S.
• Add to S all minimal generalizations h of s such that
h consistent with d and Some member of G is
more general than h
• Remove from S any hypothesis that is more general than
another hypothesis in S
43. Candidate-Elimination Algorithm (detailed)
G maximally general hypothesis in H
S maximally specific hypothesis in H
For each training example d = <x , c ( x )>
If d is a negative example
•Remove from S any hypothesis that is inconsistent with d
•For each hypothesis g in G that is not consistent with d
•Remove g from G.
•Add to G all minimal specializations h of g such that
h consistent with d and Some member of S is
more specific than h
•Remove from G any hypothesis that is less general than
another hypothesis in G
44. Candidate-Elimination Algorithm (detailed)
G maximally general hypothesis in H
S maximally specific hypothesis in H
For each training example d = <x , c ( x )>
If d is a negative example
•Remove from S any hypothesis that is inconsistent with d
•For each hypothesis g in G that is not consistent with d
•Remove g from G.
•Add to G all minimal specializations h of g
such that
h consistent with d and Some
member of S is more specific
than h
•Remove from G any hypothesis that is less
general than another hypothesis in G
If d is a positive example
• Remove from G any hypothesis that is inconsistent with d
• For each hypothesis s in S that is not consistent with d
• Remove s from S.
• Add to S all minimal generalizations h of s
such that
h consistent with d and Some
member of G is more general
than h
• Remove from S any hypothesis that is more
general than another hypothesis in S
45. Problem 1 : Apply the Candidate Elimination Algorithm to obtain the final version space for the
training examples
6 attributes (Nominal-valued (symbolic) attributes):
Sky (SUNNY, RAİNY, CLOUDY),
Temp (WARM,COLD),
Humidity (NORMAL, HIGH),
Wind (STRONG, WEAK),
Water (WARM, COOL),
Forecast (SAME, CHANGE)
Sky AirTemp Humidity Wind Water Forecast EnjoySport
Sunny Warm Normal Strong Warm Same Yes
Sunny Warm High Strong Warm Same Yes
Rainy Cold High Strong Warm Change No
Sunny Warm High Strong Cool Change Yes
66. Problem 2 : Apply the Candidate Elimination Algorithm to obtain the final version space for the
training examples
67. Problem 3 : Apply the Candidate Elimination Algorithm to obtain the final version space for the
training examples
Sky
Air
Temp Humidity Wind Water Forecast Enjoy Sport
Rainy Cold High Strong Warm Change No
Sunny Warm High Strong Warm Same Yes
Sunny Warm High Strong Cool Change Yes
Sunny Warm Normal Strong Warm Same Yes
68. Problem 4 : Apply the Candidate Elimination Algorithm to obtain the final version space for the
training examples
Size Color Shape Class/Label
Big Red Circle No
Small Red Triangle No
Small Red Circle Yes
Big Blue Circle No
Small Blue Circle Yes
69. Problem 5 : Apply the Candidate Elimination Algorithm to obtain the final version space for the
training examples
Example Citations Size InLibrary Price Editions Buy
1 Some Small No Affordable One No
2 Many Big No Expensive Many Yes
3 Many Medium No Expensive Few Yes
4 Many Small No Affordable Many Yes
70. Problem 6 : Apply the Candidate Elimination Algorithm to obtain the final version space for the
training examples
71. Problem 7 : Apply the Candidate Elimination Algorithm to obtain the final version space for the
training examples