Variosalgoritmos - Jupyter Notebook
Variosalgoritmos - Jupyter Notebook
REGRESION LINEAL
In [28]: import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
In [29]: df = pd.read_csv("Advertising2.csv")
In [100]: df.tail()
In [39]: modelo.fit(X_train,y_train)
Out[39]: LinearRegression()
In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook.
On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.
In [40]: modelo.intercept_
Out[40]: 3.1515267680706494
In [41]: modelo.coef_
Out[44]: 37 14.7
109 19.8
31 11.9
89 16.7
66 9.5
Name: sales, dtype: float64
In [47]: MAE
Out[47]: 1.2137457736144808
In [48]: MSE
Out[48]: 2.298716697886378
In [49]: RMSE
Out[49]: 1.5161519375993877
In [50]: df['sales'].mean()
Out[50]: 14.0225
REGRESION POLINOMIAL
In [90]: ejecutar_modelo(canal, X_train, y_train, X_test, y_test )
RMSE: 1.2580
In [51]: modelPol=LinearRegression()
RMSE: 1.2580
RMSE: 0.5803
RMSE: 1.2580
KNN
In [95]: from sklearn.neighbors import KNeighborsRegressor
In [96]: valores_k = [1,5,10,15]
for n in valores_k:
modelo_knn = KNeighborsRegressor(n_neighbors=n)
ejecutar_modelo(modelo_knn,X_train, y_train, X_test, y_test)
RMSE: 1.6936
RMSE: 1.5412
RMSE: 1.9077
RMSE: 2.2057
ARBOLES DE DECISIÓN
In [97]: from sklearn.tree import DecisionTreeRegressor
modelo_arbolDeci=DecisionTreeRegressor()
RMSE: 1.0120
In [99]: modelo_arbolDeci.get_n_leaves()
Out[99]: 132
Conclusiones
Debido a los parametros de evaluación se puede concluir que el mejor algoritmo de predicción para este caso es la regre
sión polinomial en grado 3, para el algoritmo de bosques aleatorios la escazes de datos hace insignificante su uso, ya
que con el arbol de decisión se obtienen mejores resultados para este dataset reducido
In [ ]: