Why Adjusted-R Square Test:
R-square test is used to determine the goodness of fit in regression analysis. Goodness of fit implies how better regression model is fitted to the data points. More is the value of r-square near to 1, better is the model. But the problem lies in the fact that the value of r-square always increases as new variables(attributes) are added to the model, no matter that the newly added attributes have a positive impact on the model or not. also, it can lead to over fitting of the model if there are large no. of variables.
Adjusted r-square is a modified form of r-square whose value increases if new predictors tend to improve model’s performance and decreases if new predictors does not improve performance as expected.
For better understanding consider :
Average Fitted Line
Best Fitted Line :
Clearly, SStot is always fixed for some data points if new predictors are added to the model, but value of SSres decreases as model tries to find some correlations from the added predictors. Hence, r-square’s value always increases.
Adjusted R-Square :
Here, k is the no. of regressors and n is the sample size.
if the newly added variable is good enough to improve model’s performance, then it will overwhelm the decrease due to k. Otherwise, increase in k will decrease adjusted r-square value.
R-square value and adjusted r-square value 0.957, 0.955 respectively. But when an attribute
Id is added, which is an irrelevant attribute, gives r-square and adjusted r-square equal to
0.958, 0.954 respectively.
Hence on adding an irrelevant attribute in dataset, the value of r-square increases(from 0.957 to 0.958). But value of adjusted r-square decreases(from 0.955 to 0.954).
- Heteroscedasticity in Regression Analysis
- ML | R-squared in Regression Analysis
- ML | Linear Regression vs Logistic Regression
- ML | Linear Regression
- ML - Different Regression types
- ML | Classification vs Regression
- Simple Linear-Regression using R
- ML | Logistic Regression using Tensorflow
- Linear Regression using PyTorch
- Types of Regression Techniques
- Seaborn | Regression Plots
- Logistic Regression using Statsmodels
- Multiple Linear Regression using R
- ML | Logistic Regression using Python
- Non linear Regression examples - ML
- Understanding Logistic Regression
- Least Angle Regression (LARS)
- Logistic Regression in R Programming
- ML | Why Logistic Regression in Classification ?
- Linear Regression Using Tensorflow
If you like GeeksforGeeks and would like to contribute, you can also write an article using contribute.geeksforgeeks.org or mail your article to email@example.com. See your article appearing on the GeeksforGeeks main page and help other Geeks.
Please Improve this article if you find anything incorrect by clicking on the "Improve Article" button below.
Improved By : shubham_singh