# Gradient Descent in Linear Regression

In linear regression, the model targets to get the best-fit regression line to predict the value of y based on the given input value (x). While training the model, the model calculates the cost function which measures the Root Mean Squared error between the predicted value (pred) and true value (y). The model targets to minimize the cost function.

To minimize the cost function, the model needs to have the best value of θ_{1} and θ_{2}. Initially model selects θ_{1} and θ_{2} values randomly and then itertively update these value in order to minimize the cost function untill it reaches the minimum. By the time model achieves the minimum cost function, it will have the best θ_{1} and θ_{2} values. Using these finally updated values of θ_{1} and θ_{2} in the hypothesis equation of linear equation, model predicts the value of x in the best manner it can.

Therefore, the question arises – **How θ _{1} and θ_{2} values get updated ?**

**Linear Regression Cost Function:**

**Gradient Descent Algorithm For Linear Regression**

->θWeights of the hypothesis. ->_{j}:hpredicted y value for i_{θ}(x_{i}) :^{th}input. ->j :Feature index number (can be 0, 1, 2, ......, n). ->α :Learning Rate of Gradient Descent.

We graph cost function as a function of parameter estimates i.e. parameter range of our hypothesis function and the cost resulting from selecting a particular set of parameters. We move downward towards pits in the graph, to find the minimum value. Way to do this is taking derivative of cost function as explained in the above figure. Gradient Descent step downs the cost function in the direction of the steepest descent. Size of each step is determined by parameter **α** known as **Learning Rate**.

In the Gradient Descent algorithm, one can infer two points :

**If slope is +ve**: θ_{j}= θ_{j}– (+ve value). Hence value of θ_{j}decreases.**If slope is -ve**: θ_{j}= θ_{j}– (-ve value). Hence value of θ_{j}increases.

The choice of correct learning rate is very important as it ensures that Gradient Descent converges in a reasonable time. :

- If we choose
**α to be very large**, Gradient Descent can overshoot the minimum. It may fail to converge or even diverge.

- If we choose α to be very small, Gradient Descent will take small steps to reach local minima and will take a longer time to reach minima.

- ML | Stochastic Gradient Descent (SGD)
- Optimization techniques for Gradient Descent
- Gradient Descent algorithm and its variants
- ML | Mini-Batch Gradient Descent with Python
- ML | Linear Regression
- Multiple Linear Regression using R
- Simple Linear-Regression using R
- Linear Regression using PyTorch
- Linear Regression Using Tensorflow
- ML | Rainfall prediction using Linear regression
- ML | Multiple Linear Regression using Python
- Python | Linear Regression using sklearn
- ML | Normal Equation in Linear Regression
- Linear Regression (Python Implementation)
- Univariate Linear Regression in Python

For linear regression Cost Function graph is always convex shaped.

**Note: ** Gradient descent sometimes is also implemented using Regularization.

## Recommended Posts:

If you like GeeksforGeeks and would like to contribute, you can also write an article using contribute.geeksforgeeks.org or mail your article to contribute@geeksforgeeks.org. See your article appearing on the GeeksforGeeks main page and help other Geeks.

Please Improve this article if you find anything incorrect by clicking on the "Improve Article" button below.