ML End-Sem
ML End-Sem
Unsupervised learning
It is a type of machine learning where algorithms are used to uncover patterns or hidden
structures in unlabeled data. Unlike supervised learning, where the algorithm learns from labeled
data (input-output pairs), unsupervised learning deals with input data that doesn't have
corresponding output labels.
There are several approaches to unsupervised learning, each serving different purposes:
1. Clustering: Clustering algorithms aim to partition data points into groups or clusters based on
similarities in their features. Some popular clustering algorithms include:
• K-means: Divides data into K clusters, where each data point belongs to the cluster with the
nearest mean.
• Hierarchical clustering: Builds a hierarchy of clusters by either merging or splitting them
based on distance metrics.
• DBSCAN: Density-Based Spatial Clustering of Applications with Noise identifies clusters in
high-density areas separated by low-density regions.
Applications:
• Market Basket Analysis: Understanding which items are frequently bought together to
drive product placement, marketing strategies, or bundle offerings.
• Recommendation Systems: Generating recommendations by analyzing user-item
interactions and suggesting items based on co-occurrence patterns.
Tabular difference
Ensemble methods
Ensemble methods in machine learning refer to techniques that combine predictions from
multiple individual models to produce a stronger, more accurate predictive model. These methods
aim to improve the overall performance and robustness compared to using a single model.
Reinforcement Learning (RL)
It is a type of machine learning paradigm where an agent learns to make sequential decisions by
interacting with an environment to achieve a specific goal. In RL, the agent learns through a trial-
and-error process by receiving feedback in the form of rewards or penalties based on its actions.
Key components of reinforcement learning:
1. Agent: The learner or decision-maker that interacts with the environment. It observes the
environment, takes actions, and receives feedback.
2. Environment: The external system with which the agent interacts. It responds to the actions
taken by the agent and provides feedback in the form of rewards or penalties.
3. Actions: Choices made by the agent that influence the state of the environment.
4. State: Represents the current situation or configuration of the environment, which the agent
perceives before taking actions.
5. Rewards: Feedback signals provided by the environment to the agent after each action.
Rewards guide the agent toward maximizing cumulative reward over time, aligning with its
goal.
6. Policy: The strategy or set of rules that the agent uses to decide actions in different states.
7. Value Function: Estimates the expected cumulative reward an agent can obtain from a
particular state or action, helping the agent make better decisions.
8. Learning Process: The agent learns by interacting with the environment, using experiences
(state, action, reward) to update its policy or value function to make better decisions over
time.
9. Exploration vs. Exploitation: Balancing between exploring new actions and exploiting known
actions to maximize rewards while learning.
2. Types of TD Learning:
• SARSA (State-Action-Reward-State-Action): TD learning algorithm that updates value
estimates based on the current state-action pair and the action taken next (on-policy
method).
• Q-learning: TD learning algorithm that updates value estimates based on the current state
and the action that maximizes the value of the next state (off-policy method).