Model Performance
Check the model performance with a complex predictor matrix using regularization.
Moving forward, let's check how the penalties affect the performance of our models. Using the (X
, y
) dataset, we have not seen many benefits of regularization. We can try the second dataset (X_overfit
, y_overfit
) and see the effect of all three types of regularization and compare the results to learn how regularization helps us control overfitting. Let's start with linear regression without regularization.
Press + to interact
from sklearn.linear_model import LinearRegressionfrom sklearn.model_selection import cross_val_scorelr_model = LinearRegression()lr_cv_mean_mse = -cross_val_score(estimator=lr_model, X=X_overfit, y=y_overfit,cv=5, scoring='neg_mean_squared_error').mean()lr_cv_mean_r2 = cross_val_score(estimator=lr_model, X=X_overfit, y=y_overfit,cv=5, scoring='r2').mean()print("These are results from linear regression (cv=5) without regularization:")print("The Ridge CV mean MSE: ",lr_cv_mean_mse)print("The Ridge CV mean R^2: ", lr_cv_mean_r2)
Now, let’s use ridge instead of linear regression.
Ridge regression
Let's use
Access this course and 1400+ top-rated courses and projects.