J1(SkillDzire)
J1(SkillDzire)
MACHINE LEARNING
INTERNSHIP
An Internship Report Submitted at the end of seventh semester
BACHELOR OF TECHNOLOGY
IN
COMPUTER SCIENCE AND ENGINEERING
Submitted By
Y SAI PRAJWAL
(21981A05J1)
Under the esteemed guidance of
Srikanth Muppala
The results embodied in the internship report have not been submitted to any
other
University or Institute for the award of any Degree.
INTERNAL GUIDE HEAD OF DEPARTMENT
P Mounika, Dr. R.Sivaranjani
Assistant Professor Professor
Dept of CSE, Dept of CSE,
Raghu Engineering College Raghu Engineering College
Dakamarri(v),Visakhapatnam. Dakamarri(v),Visakhapatnam.
EXTERNAL EXAMINER
I
DISSERTATION APPROVAL
BY
Y SAI PRAJWAL
(21981A05J1)
Is approved for the degree of Bachelor of
Technology
P Mounika
Project Guide
(Assistant Professor)
Internal Examiner
External Examiner
Dr.R.Sivaranjani
HOD
(Professor)
Date:
II
DECLARATION
This is to certify that this internship titled "Machine Learning" is bonafide work done by
me, impartial fulfilment of the requirements for the award of the degree B.Tech
and submitted to the Department of Computer Science and Engineering,
Raghu Engineering College, Dakamarri.
I also declare that this internship is a result of my own effort and that has not been copied
from anyone and I have taken only citations from the sources which are mentioned in the
references.
This work was not submitted earlier at any other University or Institute for the reward of
any degree.
Place: (21981A05J1)
I
CERTIFICATE
IV
ACKNOWLEDGEMENT
I express sincere gratitude to my esteemed Institute "Raghu Engineering
College",
which has provided us an opportunity to fulfill the most cherished desire to reach my goal.
I take this opportunity with great pleasure to put on record our ineffable
personal indebtedness to Mr. Raghu Kalidindi, Chairman of Raghu Engineering
College for providing necessary departmental facilities.
I sincerely express our deep sense of gratitude to Dr.R. Sivaranjani, Professor, Head of
Department in Department of Computer Science and Engineering, Raghu Engineering
College, for her perspicacity, wisdom and sagacity coupled with compassion and patience.
It is my great pleasure to submit this work under her wing. I thank you for guiding us for
the successful completion of this project work.
I would like to thank the team of Srikanth Muppala for providing the technical guidance
to carry out the module assigned. Your expertise in the subject matter and
dedication towards our project have been a source of inspiration for all of us.
I extend my deep hearted thanks to all faculty members of the Computer Science
department for their value based imparting of theory and practical subjects, which were
used in the project.
Regards
Y SAI PRAJWAL
(21981A05J1)
V
TABLE OF CONTENTS
1. Introduction 1
6. Module-5: Introduction to ML
15
7. Module-6: Applications of ML
19
8. Module-7: Types of ML
23
9. Module-8: Computer Vision,CNN
27
10. Module-9:Algorithms(YOLO)
31
11. Module-10:NLP processing
33
12.
Annexure
38
13.
Conclusion 41
INTRODUCTION
Machine learning (ML) is a subset of artificial intelligence (AI) focused on
developing algorithms that enable computers to learn from and make decisions based on
data. Unlike traditional programming, where a system follows explicit instructions,
machine learning allows systems to identify patterns, adapt to new data, and improve
performance over time without being explicitly programmed.
The need for machine learning arises from the exponential growth of data and the
complexity of extracting meaningful insights from it. In today's data driven world,
organisations generate vast amounts of information from various sources, such as social
media, sensors, and transaction logs. Machine learning algorithms can process and analyse
this data to uncover hidden patterns, predict future trends, and optimise decision making
processes.
By leveraging machine learning, industries ranging from healthcare to finance can achieve
significant advancements. For instance, ML models can enhance diagnostic accuracy
in medical imaging, detect fraudulent activities in financial transactions, and
personalise customer experiences in ecommerce. As a result, machine learning is
increasingly becoming a critical tool for innovation and competitiveness in the modern
technological landscape.
1
MODULE 1: INTRODUCTION TO PYTHON
OBJECTIVES:
1.Introduction to python
3.Applications of Python
4.Conditional statements
5.Python functions
1.Lambda function
2. Kwargs function
INTRODUCTION TO PYTHON:
Python is an interpreted, highlevel, dynamically typed programming language known for
its emphasis on code readability and simplicity. It features a multi paradigm
approach, supporting procedural, object oriented, and functional programming styles.
Python is implemented using CPython, which is written in C, and it utilises
automatic memory management and a builtin garbage collector. The language
provides extensive standard libraries and frameworks, making it suitable for a wide range
of applications, from web development to data analysis and artificial intelligence. Python
adheres to the principles outlined in the Zen of Python, promoting simplicity and clarity in
coding practices.
The vibrant community contributes to a rich ecosystem of third party libraries and
frameworks, particularly for data science and machine learning. Python's cross platform
compatibility and integration capabilities make it suitable for diverse environments
and applications.
2
DATA TYPES IN
PYTHON
1.Integers/Floating point numbers(Real numbers)/Complex numbers
2.Strings
3.Lists
4.Dictionary
5.Sets(Maps)
For the development of the machine learning models we will be particularly using the lists,
dictionaries and sets
Dictionary: It’s a pairwise collection of data, where the there is key value pairs ex:
dict={“hi”:”hello”,1:”next”}, here keys need to be unique and can be of any data type
Lists: it’s a linear sequential collection of data of any data type,which have
methods associated to it, where it can be iterated, sorted, appended , l=[2,3,”char”]
Sets: It’s a data type where it’s possible to have only unique values,we can perform all the
set operations of union,intersection and difference on two sets. No duplicate values
are tolerated, and the values are not sequentially arranged in memory.
Applications of Python: With its ability to perform cross platform python is used to create
web applications as well as to analyse data and work on the machine learning and
furthermore.
Functions: Python functions, take parameters and work on the logic, Functions are
initiated only through function calls making it modular for operations, with discreet
allocation of memory, it’s functional for operation
Lambda Functions: These are special functions,these are anonymous functions where the
logic is distributed to the iterable parameter of the lambda functions
3
MODULE 2: OOPS IN PYTHON
Objective:
1. OOPs concept in Python
● Pillars of OOPs and examples
ObjectOriented Programming (OOP) in Python is a programming paradigm
centred around objects, which encapsulate data and behaviour. The four fundamental
principles of OOP are:
1.Encapsulation: Bundling data (attributes) and methods (functions) that operate on the
data into a single unit, or object. This promotes data hiding and protects the integrity of the
object's state.
2.Abstraction: Simplifying complex systems by exposing only the relevant attributes and
methods, allowing users to interact with objects without needing to understand their
internal workings.
3.Inheritance: Enabling a new class (subclass) to inherit attributes and methods from an
existing class (superclass), promoting code reuse and hierarchical_relationships.
In Python, classes are defined using the `class` keyword, and objects are instantiated from
these classes. OOP facilitates modularity, code reusability, and easier maintenance in
software development.
OOPs concepts can be implemented as the following code below, where we class
has
functions(methods) and data(attributes) associated with it.
4
Encapsulation: Encapsulation in Python, is the feature of hiding classes(objects) and its
attributes and methods within other classes, this controls the access of the attributes and
methods outside the class.
5
Inheritance:Inheritance in Python allows a class to inherit attributes and methods from
another class, promoting code reuse and establishing a hierarchical relationship between
classes. Here’s an overview of the types of inheritance possible in Python.
Single or Simple Inheritance: This is when there is a single Parent class the Child class
inherits, this can be demonstrated as the following.
6
Abstraction:Abstraction concept is the idea of having a template,blueprint for
classes which could be reused to create more classes where the abstracted base class is
inherited to access abstract methods, in Python we have to import the library
abc(Abstract Base Class), Abstraction can be explained as following example.
Area:50 Area:154
Shape is abstract base class and it’s inherited by Rectangle class and Circle class, and the
abstract methods area and perimeter are appended.
7
Polymorphism: Polymorphism in Python allows different classes to be treated as
instances of the same class through a common interface. It typically manifests in two ways:
through method overriding and method overloading. However, Python does not support
traditional method overloading (as seen in some other languages), but you can
achieve similar functionality using default arguments or variable arguments.
1.Method overloading
2.Method overriding
3.Polymorphism in collections
8
Method overriding: This method is when we are going to change the logic of the method
that is being inherited without changing the name of the method, Here the Animal class’s
method sound is overridden by class Dog and Cat
9
MODULE 3: PANDAS,NUMPY LIBRARIES INTRODUCTION
Objective:
1. Introduction to Numpy
2. Introduction to Pandas
3. Applications of these Libraries
NumPy
Introduction:
Data Manipulation: NumPy arrays are often used to represent datasets, enabling
efficient mathematical computations.
1
The demonstration of the this library can be seen here
Pandas
Introduction:
Pandas is a powerful data manipulation and analysis library that provides data
structures like Series and DataFrames, which are built on top of NumPy.
Data Handling: Pandas is often used for data loading, cleaning, and preparation,
which are critical steps in the machine learning workflow.
1
Feature Engineering: Pandas makes it easy to create, transform, and manipulate
features that can be used as inputs to machine learning algorithms.
While NumPy is excellent for numerical computing and is very efficient for operations on
homogeneous data, Pandas is designed for data analysis and manipulation, especially with
heterogeneous, labelled data. If you're working with structured data (like tables) or
performing data analysis tasks, Pandas is often the more convenient and powerful choice.
In practice, you might use both libraries together: leveraging NumPy for numerical
computations and Pandas for data manipulation and analysis.
1
MODULE 4: DATAANALYSIS
Objective:
1.Function
2.Formula
3.Charts
4.Pivots
Data analysis is an important primary task to understand basic statistical parameters which
could give us attributes and parameters to consider to predict the movement of the overall
data in any period of time.
1.SUM
2.AVG
3.MAX
4.CONCATENATE
Formula:
In Excel, a formula is an expression that performs calculations on values in your
spreadsheet. Formulas can include functions, operators, cell references, and
constants. They allow you to perform a wide variety of mathematical and logical
operations on your data.
1
The following are the basic demonstration of formula in Excel
Charts:
Charts in Excel are powerful tools for visually representing data, making it easier
to identify trends, patterns, and comparisons. Excel offers a variety of chart types, each
suited for different types of data analysis. The types of charts:
1.Column Chart
2.Bar chart
3.Pie Chart
4.Line Chart
5.Scatter Plot
Excel acts as a primary software tool which could help with the raw data and with it’s
manipulation of data using some of the prominent features mentioned above useful
inferences can be reproduced, to predict the possibility of the near future and that
prediction falls under the category of machine learning which is unravelled in the next
module
1
MODULE 5: INTRODUCTION TO MACHINE LEARNING
Objective:
1.Introduction to machine learning
Machine learning:
Depending on the outcomes and data, machine learning is divided into the
following categories
1.Supervised Learning The model is trained on labelled data, meaning the input data is
paired with the correct output.
2.Unsupervised Learning: The model is trained on unlabeled data and must find patterns
or structures within the data on its own.
Examples:
1
Training agents to play games (e.g., AlphaGo).
Robotics, where a robot learns to navigate and perform tasks through trial and error.
These types of machine learning cater to different kinds of problems and datasets, each
with its unique methodologies and applications.
1957: Frank Rosenblatt develops the Perceptron, an early neural network model.
Achievements in image and speech recognition, natural language processing, and game
playing AI.
1
Machine learning becomes mainstream across various sectors, including healthcare,
finance, and autonomous systems.
Why Python is chosen and what are the libraries and toolkits that are used:
Python is generally much preferred because of its readability of the code and its simplicity
makes it user friendly and also makes it less intimidating to new learners, and its strong
community keeps this language relevant to this field.
Libraries which are used here in this are NumPy and Pandas which are already discussed in
the module 3.MatPlotLib and Scikit are libraries of python where we are using it for data
visualization.
Frameworks like TensorFlow and PyTorch:
Flexibility: TensorFlow supports various machine learning tasks, including supervised and
unsupervised learning, reinforcement learning.
HighLevel APIs: TensorFlow includes high level APIs like Keras, which make it easier to
build and train neural networks without dealing with low level details.
Scalability: It can run on various platforms, from mobile devices to large scale distributed
systems, making it suitable for both research and production environments.
1
PyTorch: PyTorch is an open source machine learning library developed by Facebook’s
AI Research lab. It’s particularly popular for its dynamic computation graph and ease of
use, making it a favourite among researchers and practitioners in deep learning. Here are
some of its key features and uses:
1. Dynamic Computation Graphs: Allows for changes to be made on the fly during
model training, making it easier to debug and experiment.
2. Easy to Use: PyTorch's syntax is intuitive and resembles standard Python, making
it accessible for beginners.
3. Rich Ecosystem: Supports a wide range of libraries for tasks like computer vision
(TorchVision) and natural language processing (TorchText).
4. Strong Community Support: A growing community that contributes to a variety of
tutorials, resources, and libraries.
Though both the frameworks serve similar purposes, these are different in the functionality
and the learning curves. Trends of this Frameworks are as the following in the past 5 years
1
1.Machine Learning Responsibilities
4.Popular Algorithms in ML
These are the results and responsibilities that are expected of Machine Learning
1. Improved DecisionMaking
2. Automation of Tasks
3. Personalization
4. Predictive Analytics
5. Advancements in Healthcare
7. Autonomous Systems
9.Enhanced Security
How are the outputs of Machine Learning interpreted and can be used to make an inference
on the subject.Machine learning outputs can be conveyed in various forms,
including numbers, graphs, and charts. Here’s how these outputs typically look and how to
interpret them:
1. Numerical Outputs
Predictions: For regression tasks, you might receive a single numerical value, such as a
predicted price (e.g., $250,000 for a house).
1
Classification Probabilities: For classification tasks, outputs may look like:
`Class A: 0.7`
`Class B: 0.3`
This means there’s a 70% probability that the instance belongs to Class A.
For regression, the predicted value can be directly compared to actual values to
assess accuracy.
For classification probabilities, you can set a threshold (e.g., 0.5) to decide which class to
assign.
Scatter Plots: Used for regression analysis, displaying actual vs. predicted values.
The closer points are to the diagonal line, the better the model performs. A good fit will
show points clustered near the diagonal, indicating accurate predictions.
Confusion Matrix: For classification problems, a confusion matrix displays true positives,
true negatives, false positives, and false negatives.
From the confusion matrix, you can derive metrics like accuracy, precision, recall, and F1
score:
3. Heatmaps
1. Automation: It automates tasks that typically require human intelligence, like data
analysis and decision making.
2. Personalization: Used in recommendation systems (like Netflix or Amazon)
to tailor content and products to individual users.
3. Predictive Analytics: Helps in forecasting trends, such as sales predictions, stock
market trends, and customer behaviour.
4. Image and Speech Recognition: Powers technologies in facial recognition, voice
assistants (like Siri and Alexa), and autonomous vehicles.
5. Healthcare: Assists in diagnosing diseases, predicting patient outcomes, and
personalizing treatment plans.
6. Fraud Detection: Identifies unusual patterns and behaviours in financial
transactions, helping to combat fraud.
7. Natural Language Processing: Enables applications like chatbots, translation
services, and sentiment analysis.
2
Fig-2: 7 stages of Machine Learning
2. Logistic Regression
3. Decision Tree
4. SVM
5. Naive Bayes
6. KNN
7. K-Means
8. Random Forest
2
Objective:
1.Introduction to types of ML
2.Linear Regression
3.Decision Tree
4.Neural Networks
5.Clustering
Supervised Learning: This is the type of learning where there are labelled inputs mapping
to labelled defined outputs, where we know all the possible co-domains of the knowledge.
Supervised learning provides a singular output for the model trained under it.
Classification:These algorithms are used when the output variable is categorical (discrete).
The goal is to assign inputs to one of several predefined categories.
Regression:These algorithms are used when the output variable is continuous. The goal is
to predict a numerical value.
Decision Trees: Tree-like models that make decisions based on feature values.
Random Forest: An ensemble of decision trees that improves accuracy and reduces
overfitting.
K-Nearest Neighbors (KNN): Classifies based on the majority class among the k
nearest neighbours in the feature space.
These techniques reduce the number of features in a dataset while preserving its essential
characteristics.
This technique identifies relationships between variables in large datasets, often used in
market basket analysis. Common algorithms include:
Apriori Algorithm: Finds frequent itemsets and generates association rules based
on support and confidence.
Eclat Algorithm: A more efficient algorithm for finding frequent itemsets that
uses a depth-first search strategy.
4. Anomaly Detection
Anomaly detection algorithms identify outliers or unusual patterns in data. These can be
useful in fraud detection, network security, and quality control. Techniques include:
These are a type of artificial neural network that uses unsupervised learning to produce a
low-dimensional representation of the input space, helping to visualize and understand the
structure of high-dimensional data.
Linear Regression: Linear regression is one of the simplest and most commonly
used algorithms in machine learning for predictive modelling. It is used to establish
a relationship between one or more independent variables (features) and a
continuous dependent variable (target). Here’s a comprehensive overview:
2
The basic idea of linear regression is to find a linear relationship between the input features
and the target variable.
y=β0+β1x1+β2x2+…+βnxn +ϵ
Here y is the predicted dependent variable and xi, i>0 are the independent variables.
Sample Project: This is to predict the house price prediction using linear regression
Libraries:NumPy,Pandas, MatPlotlib,sklearn,seaborn
Output: we run test data test.csv to check the prediction accuracy of the linear regression.
Neural Networks: Neural networks are a class of machine learning models inspired by the
structure and function of the human brain. They are designed to recognize patterns and
make decisions based on input data.
3,Transformers
2
Objective:
2
1.Computer Vision
2.CNN
3.Process CNN
Computer Vision:
Key Features:
2
A CNN typically consists of several layers, each designed to perform specific functions.
The main components of a CNN include:
1. Convolutional Layer:
○ Function: Applies convolution operations to the input data using filters
(kernels). Each filter detects specific features, such as edges or textures.
○ Output: Produces feature maps that represent the presence of certain
features in the input.
2. Activation Function:
○ Commonly uses the Rectified Linear Unit (ReLU) activation function, which
introduces non-linearity to the model.
3. Pooling Layer:
○ Function: Reduces the dimensionality of the feature maps (downsampling)
while retaining important information. Common pooling methods
include max pooling and average pooling.
○ Benefit: Helps to make the network invariant to small translations in
the input.
4. Fully Connected Layer:
○ After several convolutional and pooling layers, the network typically flattens
the feature maps and connects them to a fully connected layer, where each
neuron is connected to all neurons in the previous layer.
○ Function: Makes the final classification based on the learned features.
5. Output Layer:
○ Produces the final predictions, often using a softmax function for multi-class
classification tasks.
2
Fig 3: Demonstration of working CNN
CNN Algorithms: While the structure of CNNs can vary, the most common
training algorithm used is backpropagation in conjunction with gradient descent. Here's
a brief outline of the process:
1. Forward Pass: Input data is passed through the layers, and the output is computed.
2. Loss Calculation: The output is compared to the actual label using a loss function
(e.g., cross-entropy loss).
3. Backward Pass: The gradients of the loss with respect to the weights are
computed using backpropagation.
4. Weight Update: The weights are updated based on the computed gradients and a
learning rate, typically using optimization algorithms like Stochastic Gradient
Descent (SGD) or Adam.
Application of CNN:
Object Detection: Locating and identifying objects within an image (e.g., bounding boxes
around detected objects).
3
Image Segmentation: Dividing an image into segments to classify each part (e.g.,
identifying different regions in a medical image).
Computer vision is integration of hardware and software solutions to real world problems
using the CNNs as the logic to achieve the goals of recognition and image processing. In a
way CNNs are subsets of Computer Vision.
3
MODULE 9: REAL TIME OBJECT DETECTION USING
YOLO,YOLOV3 ALGORITHM
Objective:
3.Process of detection
Computer vision application: Real time of detection and recognition are quite valuable in
the aspect of business applications. These are the applications of computer vision. To
achieve this there are certain algorithms to recognize/detect. The efficiency of computer
vision is the latency it takes to recognise and detect.
Traditional Methods
Haar Cascades: A machine learning object detection method that uses a cascade of
classifiers to identify objects based on features derived from Haar-like
features. Commonly used for face detection.
Sliding Window: A method that involves moving a fixed-size window across the
image to detect objects at different scales. It is computationally expensive and has
largely been replaced by more efficient methods.
YOLO (You Only Look Once): A real-time object detection system that frames
detection as a single regression problem, predicting bounding boxes and class
probabilities directly from images. It is known for its speed and efficiency.
3
Bounding Box: coordinates (usually in the form of x, y, width, and height) and is drawn
around an object of interest in an image. It indicates the position and extent of the object.
YOLO(You only look once): Prior detection systems repurpose classifiers or localizers to
perform detection. They apply the model to an image at multiple locations and scales. High
scoring regions of the image are considered detections. YOLO uses a totally
different approach.
Apply a single neural network to the full image. This network divides the image
into regions and predicts bounding boxes and probabilities for each region. These
bounding boxes are weighted by the predicted probabilities.It looks at the whole image at
test time so its predictions are informed by global context in the image.
YOLO uses the Darknet network. The YOLO network splits the input image into a grid of
SxS cells.
Each grid cell predicts B number of bounding boxes and their objectness score along with
their class predictions. Coordinates of B bounding boxes-YOLO predicts 4 coordinates for
each bounding box (bx,by,bw,bh) with respect to the corresponding grid cell.
Objectness Score(P0) indicates the probability of the presence of an object in the given
grid(rectangle). YOLO is quite fast to realize the objects in the captured images. Versions
of YOLO algorithms are used in self-driving cars, back cameras to detect the space for
parking.
3
MODULE 10: NATURAL LANGUAGE PROCESSING (NLP)
Objective:
1.Introduction to NLP
2.Subsets of NLP
3.Algorithms of NLP
4.Working of NLP
NLP can be expressed as the two technologies, where the common language is divided into
structured logical relationships among the sentences which could help the computer predict
the sentences and improve its model. The ability to understand natural language paves a
path to more inventions such as chat bots, voice recognition and voice assistants.
3
Fig-7: Demonstration of NLP
Supervised Learning: Used for tasks where labelled data is available (e.g., text
classification, sentiment analysis).
3
4. Contextual Understanding: Grasping the context in which words or phrases are
used to derive meaning, often involving disambiguation of words with
multiple meanings.
5. Parsing: Analyzing the grammatical structure of sentences to understand their
components and relationships.
1. Text Generation: Creating coherent and contextually relevant text from structured
data or prompts (e.g., generating news articles, summaries, or conversational
responses).
2. Template-Based Generation: Using predefined templates to produce
structured text (e.g., generating reports or product descriptions based on specific
parameters).
3. Dialogue Generation: Crafting responses in a conversational setting, such as
in chatbots or virtual assistants, ensuring the output is contextually relevant
and engaging.
4. Content Creation: Generating creative content such as stories, poems, or
social media posts.
Feedback Loop: Insights gained from NLU can inform NLG models,
allowing them to generate more contextually relevant and accurate outputs.
3
Algorithms Used in NLP
1. Traditional Algorithms:
○ Naive Bayes: Often used for text classification and sentiment analysis.
○ Support Vector Machines (SVM): Effective for classification tasks.
○ Decision Trees: Can be used for text classification.
2. Vectorization Techniques:
○ Bag of Words (BoW): Represents text as a collection of words disregarding
grammar and order.
○ TF-IDF (Term Frequency-Inverse Document Frequency): Weighs the
importance of words based on their frequency in a document relative to
a collection of documents.
3. Deep Learning Models:
○ Recurrent Neural Networks (RNNs): Suitable for sequential data like text;
often used for language modelling and sequence generation.
○ Long Short-Term Memory (LSTM): A type of RNN that handles long-
term dependencies better, commonly used for tasks like machine translation.
○ Convolutional Neural Networks (CNNs): Can be adapted for text
classification tasks.
○ Transformers: A powerful architecture used in models like BERT and GPT
that excels at understanding context and relationships in text.
4. Pre-trained Models:
○ BERT (Bidirectional Encoder Representations from Transformers):
Designed to understand the context of words in a sentence.
○ GPT (Generative Pre-trained Transformer): Focused on text generation
and conversation.
3
Example of sentence is logically structured:“The cat sat on the mat."
ANNEXURE
3
We are to be creating a Machine Learning Linear Regression Model. You will be analysing
a house price prediction dataset for finding out the price of house on different parameters.
You will do Exploratory Data Analysis, split the training and testing data, Model
Evaluation and Predictions.
Problem Statement:A real estate agent wants the help to predict the house price
for regions in the USA. He gave you the dataset to work on and you decided to use the
Linear Regression Model. Create a model which will help him to estimate what the house
would sell for. Dataset contains 7 columns and 5000 rows with CSV extension. The data
contains the following columns :
'Avg. Area Income','Avg. Area House Age' 'Avg. Area Number of Rooms':'Avg.
Area
Number of Bedrooms': 'Area Population''Price''Address'
Libraries:NumPy,Pandas, MatPlotlib,sklearn,seaborn
3
Schema of the dataframe:
39
Splitting the data and check the effectiveness of the division through the coefficients, data
with varied information is quite helpful to test the limits of the model.
Conclusion:From the distribution of the data points which if forming the line shape
indicates the data is performing good with the linear regression model.
4
CONCLUSION:
This internship has covered all the prominent features of Machine Learning and its
applications in real life. Machine learning is a subset of Artificial Intelligence and its idea
is built on the emphasis of pattern formation and prediction which is associated with Deep
Learning. Machine Learning is applicable to any form of data, whether it’s image,
numerical data, text, audio, or video, with the mathematical relationship within this digital
media, it’s predictable to manipulate data and understand the underlying patterns. Machine
Learning is one of the earliest discovered concepts which was majorly researched in the
early 1950s to understand patterns for facial recognition to distinguish between male and
female. Machine learning will advance further in its ability to understand information and
make better blueprint of the information without violating the ethics of intelligence and
considering potential biases.