Optimizing Polynomial and Regularization Techniques for Enhanced Housing Price Prediction Accuracy

Document Type

Article

Publication Title

SN Computer Science

Abstract

This study investigates the effectiveness of various regression models for predicting housing prices using the California Housing dataset. The models evaluated include Linear Regression, Ridge Regression, Best Polynomial Ridge Regression, Lasso Regression, Elastic Net Regression, and Support Vector Regression (SVR) with an RBF kernel. The analysis reveals that SVR with an RBF kernel exhibits the poorest performance, characterized by the highest Mean Squared Error (MSE) and the lowest R² score, indicating limited effectiveness for this dataset. Conversely, Linear Regression, Ridge Regression, and Best Polynomial Ridge Regression demonstrate significantly lower MSE values and nearly identical R² scores, each explaining approximately 60% of the variance in housing prices. Among these, Best Polynomial Ridge Regression marginally outperforms Linear and Ridge Regression, suggesting that including polynomial features enhances model performance. Although Lasso Regression shows slightly higher MSE than the leading models, it still performs better than Elastic Net and SVR. Overall, the study identifies linear regression, ridge regression, and best polynomial ridge regression as the most effective models for this dataset, with SVR and an RBF kernel being the least effective.

DOI

10.1007/s42979-024-03578-7

Publication Date

2-1-2025

This document is currently not available here.

Share

COinS