Ml Lab Manual Cse
Ml Lab Manual Cse
INSTITUTE OF TECHNOLOGY
(Approved by AICTE New Delhi & Affiliated to JNTUK, Kakinada)
NH- 5, Chowdavaram, Guntur-522 019
An ISO 9001:2015 Certified Institution, Accredited by NAAC & NBA
KHIT
PRACTICAL RECORD
Name:……………………………………………………………………..
Roll No:……………..…... Year & Semester:…………..………..
Branch:…………………. Section:…………………………........
Lab:……………………………………………………………………..…
KALLAM HARANADHAREDDY
INSTITUTE OF TECHNOLOGY
(APPROVED BY AICTE NEW DELHI, AFFLIATED TO
JNTUK, KAKINADA) CHOWDAVARAM, GUNTUR-19
Roll No:
CERTIFICATE
This is to Certify that the Bonafide Record of the Laboratory Work done by
Mr/Ms…………………………………………………………………………………………..
.
EX. PAGE
NO
DATE NAME OF THE EXPERIMENT FROM TO MARKS SIGNATURE
.
CSE DEPARTMENT VISION, MISSION, GOALS
Vision
Imparting quality technical education to learners in the field of Computer Science and
Engineering to produce technically competent software personnel with advanced skills, knowledge and
behavior to meet the computational global real time challenges.
Mission
PSO-2: Understand, Analyze, Design, Develop and Test computer programs for the
problems related to Algorithms, Internet of Things, Data Sciences, Cloud Computing,
Artificial Intelligence and Machine Learning.
PSO-3: Apply theoretical and practical knowledge by using Modern software tools and
techniques to build application software.
Course Outcomes:
• Analyze and identify the need for machine learning techniques for a particular domain.
• Develop the real time applications and predict its outcomes using machine learning
algorithms.
CO- PO Mapping
(3/2/1 indicates strength of correlation) 3-Strong, 2-Medium, 1-Weak
COs Programme Outcomes(POs) PSOs
PO1 PO2 PO3 PO4 PO5 PO6 PO7 PO8 PO9 PO10 PO11 PO12 PSO1 PSO2 PSO3
CO1 1 1 1 1 1 1 2
CO2 1 1 2 1 1 1 1 1 1 1
Exp. No: Date:
1. Install the python software/Anaconda- python and install useful package for machine learning
load the dataset (sample), understand, and visualize the data.
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
import pandas as pd
import matplotlib.pyplot as plt
# importing the dataset
dataset
=pd.read_csv('D:/Salary_Data.csv')
print(dataset.head())
Output:
CSE KHIT
Exp. No: Date:
import numpy as np
def estimate_coef(x,y):
# number of observations/points
n = np.size(x)
# putting labels
plt.xlabel('x')
plt.ylabel('y')
# estimating coefficients
b = estimate_coef(x, y)
print("Estimated coefficients:\nb_0 = {}\nb_1 = {}".format(b[0], b[1]))
Output:
Estimated coefficients:
b_0 = -0.0586206896552
b_1 = 1.45747126437
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
import pandas
a={
'slips' : [2,4,6,8],
'open' : [2,4,6,8],
'marks' : [20,40,60,80]
df = pandas.DataFrame(a)
X = df[['slips', 'open']]
y = df['marks']
regr = linear_model.LinearRegression()
regr.fit(X, y)
predictedMarks= regr.predict([[5,5]])
print(predictedMarks)
Output:
[ 50 . ]
CSE KHIT
Exp. No: Date:
import pandas as pd
a={
'slips' : [2,4,6,8,10,15,18],
'pass_or_fail' : [0,1,1,1,1,1]
DF=pd.DataFrame(a)
feature_cols = ['slips']
X = DF[feature_cols] # Features
X_train,X_test,y_train,y_test=train_test_split(X,y,test_size=0.5,random_state=0)
logreg = LogisticRegression()#
logreg.fit(X_train,y_train)
y_pred=logreg.predict(X_test)
print("Accuracy:",metrics.accuracy_score(y_test, y_pred))
Exp. No: Date:
print(cnf_matrix)
Output:
runfile('C:/Users/student/.spyder-py3/temp.py', wdir='C:/Users/student/.spyder-py3')
Accuracy: 0.75
[[0 0]
[1 3]]
CSE KHIT
Exp. No: Date:
# Load libraries
import pandas as pd
'easy' : [0,1,1,0,0,0,0,0],
'slips' : [0,0,2,2,4,6,8,10],
'result' : [0,1,1,0,1,1,1,1]
pima=pd.DataFrame(a)
print(pima.head())
X = pima[feature_cols] # Features
X_train,X_test,y_train,y_test=train_test_split(X,y,test_size=0.25,random_state=1)
clf = DecisionTreeClassifier()
y_pred =clf.predict(X_test)
print("Accuracy:",metrics.accuracy_score(y_test, y_pred))
OUTPUT:
runfile('C:/Users/student/.spyder-py3/untitled1.py', wdir='C:/Users/student/.spyder-py
easy slips result
0 0 0 0
1 1 0 1
2 1 2 1
3 0 2 0
4 0 4 1
Accuracy: 1.0
CSE KHIT
Exp. No: Date:
import pandas as pd
dataset = pd.read_csv('Salary_Data.csv')
print(dataset.head())
# data preprocessing
regressor = LinearRegression()
regressor.fit(X_train, y_train)
y_pred = regressor.predict(X_test)
print(y_pred)
print(y_test)
plt.xlabel("Years of experience")
plt.ylabel("Salaries")
plt.show()
plt.xlabel("Years of experience")
plt.ylabel("Salaries")
plt.show()
Output:
CSE KHIT
Exp. No: Date:
import pandas
a={
'slips' : [2,4,6,8],
'open' : [2,4,6,8],
'marks' : [20,40,60,80]
df = pandas.DataFrame(a)
X = df[['slips', 'open']]
y = df['marks']
regr = linear_model.LinearRegression()
regr.fit(X, y)
predictedMarks= regr.predict([[5,5]])
print(predictedMarks)
Output:
[ 50 . ]
CSE KHIT
Exp. No: Date:
import pandas as pd
a={
'slips' : [2,4,6,8,10,15,18],
'pass_or_fail' : [0,1,1,1,1,1]
DF=pd.DataFrame(a)
feature_cols = ['slips']
X = DF[feature_cols] # Features
X_train,X_test,y_train,y_test=train_test_split(X,y,test_size=0.5,random_state=0)
logreg = LogisticRegression()#
logreg.fit(X_train,y_train)
y_pred=logreg.predict(X_test)
print("Accuracy:",metrics.accuracy_score(y_test, y_pred))
print("Accuracy:",metrics.accuracy_score(y_test, y_pred))
print(cnf_matrix)
Output:
runfile('C:/Users/student/.spyder-py3/temp.py', wdir='C:/Users/student/.spyder-py3')
Accuracy: 0.75
[[0 0]
[1 3]]
Exp. No: Date:
import pandas as pd
from sklearn.model_selection
import train_test_split
#Load dataset
iris = datasets.load_iris()
print(iris.target_names)
print(iris.feature_names)
print(iris.data[0:5])
print(iris.target)
data=pd.DataFrame({
'sepal length':iris.data[:,0],
'sepal width':iris.data[:,1],
'petal length':iris.data[:,2],
'petal width':iris.data[:,3],
CSE KHIT
Exp. No: Date:
'species':iris.target
})
data.head()
X=data[['sepal length', 'sepal width', 'petal length', 'petal width']] # Features y=data['species']
# Labels
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3) # 70% training and 30% test
clf=RandomForestClassifier(n_estimators=100)
clf.fit(X_train,y_train)
y_pred=clf.predict(X_test)
print("Accuracy:",metrics.accuracy_score(y_test, y_pred))
Output:
CSE KHIT
Exp. No: Date:
#Data imports
import pandas as pd
import numpy as np
#Visualization imports
plt.scatter(raw_data[0][:,0], raw_data[0][:,1])
model = KMeans(n_clusters=4)
model.fit(raw_data[0])
print(model.labels_)
print(model.cluster_centers_)
ax1.set_title('Our Model')
ax1.scatter(raw_data[0][:,0],raw_data[0][:,1],c=model.labels_)
CSE KHIT
Exp. No: Date:
ax2.set_title('Original Data')
ax2.scatter(raw_data[0][:,0], raw_data[0][:,1],c=raw_data[1])
Output:
CSE KHIT
Exp. No: Date:
#Common imports
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
#Import the data set
raw_data = pd.read_csv('classified_data.csv', index_col = 0)
print(raw_data.columns)
= StandardScaler()
scaler.fit(raw_data.drop('TARGET CLASS', axis=1))
scaled_features = scaler.transform(raw_data.drop('TARGET CLASS', axis=1))
scaled_data = pd.DataFrame(scaled_features, columns = raw_data.drop('TARGET CLASS',
axis=1).columns)
#Split the data set into training data and test data
from sklearn.model_selection import train_test_split
x = scaled_data
y = raw_data['TARGET CLASS']
x_training_data, x_test_data, y_training_data, y_test_data = train_test_split(x, y, test_size = 0.3)
#Train the model and make predictions
from sklearn.neighbors import KNeighborsClassifier
model = KNeighborsClassifier(n_neighbors = 1)
model.fit(x_training_data, y_training_data)
predictions = model.predict(x_test_data)
#Performance measurement
from sklearn.metrics import classification_report
from sklearn.metrics import confusion_matrix
CSE KHIT
Exp. No:
Date:
print(classification_report(y_test_data, predictions))
print(confusion_matrix(y_test_data, predictions))
#Selecting an optimal K value
error_rates = []
for i in np.arange(1, 101):
new_model=KNeighborsClassifier(n_neighbors=i)
new_model.fit(x_training_data,y_training_data)
new_predictions=new_model.predict(x_test_data)
error_rates.append(np.mean(new_predictions != y_test_data))
plt.figure(figsize=(16,12))
plt.plot(error_rates)
Output:
CSE KHIT
Exp. No: Date:
#Load dataset
cancer=datasets.load_breast_cancer()
# print data(feature)shape
cancer.data.shape
print(cancer.data[0:5])
print(cancer.target)
clf = svm.SVC(kernel='linear')
# Linear Kernel
CSE KHIT
Exp. No: Date:
print("Recall:",metrics.recall_score(y_test, y_pred))
Output:
CSE KHIT
Exp. No: Date:
import numpy as np
class NeuralNetwork():
np.random.seed(1)
return 1 / (1 + np.exp(-x))
return x * (1 - x)
# training the model to make accurate predictions while adjusting weights continually
output = self.think(training_inputs)
CSE KHIT
Exp. No: Date:
self.synaptic_weights += adjustments
inputs = inputs.astype(float)
return output
neural_network = NeuralNetwork()
print(neural_network.synaptic_weights)
[1, 1, 1],
[1, 0, 1],
[0, 1, 1]])
neural_network.train(training_inputs,training_outputs,15000)
print(neural_network.synaptic_weights)
CSE KHIT
Exp. No: Date:
Output:
CSE KHIT
Exp. No: Date:
import pandas as pd
import numpy as np
%matplotlib inline
data=load_breast_cancer()
data.keys()
print(data['target_names'])
print(data['feature_names'])
df1=pd.DataFrame(data['data'],columns=data['feature_names'])
scaling=StandardScaler()
scaling.fit(df1)
Scaled_data=scaling.transform(df1)
CSE KHIT
Exp. No: Date:
principal=PCA(n_components=3)
principal.fit(Scaled_data)
x=principal.transform(Scaled_data)
print(x.shape)
principal.components_
plt.figure(figsize=(10,10))
plt.scatter(x[:,0],x[:,1],c=data['target'],cmap='plasma')
plt.xlabel('pc1')
plt.ylabel('pc2')
fig = plt.figure(figsize=(10,10))
axis.scatter(x[:,0],x[:,1],x[:,2], c=data['target'],cmap='plasma')
axis.set_xlabel("PC1", fontsize=10)
axis.set_ylabel("PC2", fontsize=10)
axis.set_zlabel("PC3", fontsize=10)
CSE KHIT
Exp. No: Date:
print(principal.explained_variance_ratio_)
(569,3)
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
import pandas as pd
path = 'https://media.geeksforgeeks.org/wp-content/uploads/file.tsv'
df.head()
movie_titles=pd.read_csv('https://media.geeksforgeeks.org/wp-
content/uploads/Movie_Id_Titles.csv')
movie_titles.head()
data.head()
data.groupby('title')['rating'].mean().sort_values(ascending=False).head()
data.groupby('title')['rating'].count().sort_values(ascending=False).head()
ratings = pd.DataFrame(data.groupby('title')['rating'].mean())
ratings.head()
CSE KHIT
Exp. No: Date:
sns.set_style('white')
%matplotlib inline
ratings['rating'].hist(bins = 70)
moviemat.head()
starwars_user_ratings.head()
similar_to_starwars = moviemat.corrwith(starwars_user_ratings)
similar_to_liarliar = moviemat.corrwith(liarliar_user_ratings)
CSE KHIT
Exp. No: Date:
corr_starwars=pd.DataFrame(similar_to_starwars,columns=['Correlation'])
corr_starwars.dropna(inplace = True)
corr_starwars.head()
Output:
CSE KHIT
Exp. No: Date:
Anomaly recommendation
import pandas as pd
path = 'https://media.geeksforgeeks.org/wp-content/uploads/file.tsv'
df.head()
movie_titles = pd.read_csv('https://media.geeksforgeeks.org/wp-
content/uploads/Movie_Id_Titles.csv')
movie_titles.head()
data.head()
data.groupby('title')['rating'].mean().sort_values(ascending=False).head()
data.groupby('title')['rating'].count().sort_values(ascending=False).head()
ratings = pd.DataFrame(data.groupby('title')['rating'].mean())
ratings.head()
CSE KHIT
Exp. No: Date:
sns.set_style('white')
%matplotlib inline
ratings['rating'].hist(bins = 70)
Output:
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
CSE KHIT
Exp. No: Date:
49