-
Notifications
You must be signed in to change notification settings - Fork 0
SKT1803/SupervisedMachineLearningRegression-BlackFridaySalesPredictions
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
During data exploration and preprocessing steps, I examined the distribution of each feature, checked for missing values, and looked for potential outliers. Data cleaning involved handling missing values, removing outliers if necessary, and encoding categorical variables if applicable. Feature engineering in the later step of training included creating polynomial features to capture non-linear relationships between predictors and the target variable.
The main objective of this analysis is to develop linear regression models to predict the target variable accurately while maintaining interpretability. The focus is primarily on prediction, but it's essential to understand the relationships between the predictors and the target variable for interpretability purposes.
Four linear regression models were trained and evaluated:
Simple Linear Regression Model: This model served as the baseline and used the original features without any modifications.
Linear Regression Model with Polynomial Features Applied: Polynomial features up to the second degree were added to the original features to capture non-linear relationships.
Regularization Regression Models (Ridge and Lasso Regression): Ridge and Lasso regressions was applied to mitigate multicollinearity and overfitting by introducing regularization.
Recommendation for Final Model:
Based on the results, the polynomial regression model appears to be the most suitable final model. It achieved the lowest root mean squared error (RMSE) and a relatively higher coefficient of determination (R^2 score) compared to the other models. Moreover, the polynomial regression model captures non-linear relationships between predictors and the target variable, enhancing its predictive performance. Additionally, its coefficients can provide insights into the importance of different features in predicting the target variable.
The polynomial regression model outperforms the other models in terms of predictive accuracy.
Adding polynomial features improves the model's ability to capture complex relationships between predictors and the target variable.
Regularization techniques like Ridge and Lasso regression did not significantly improve the model's performance in this case.
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published