Skip to content

Machine Learning, EDA, Feature Engg, PLot, Transformation of features

Notifications You must be signed in to change notification settings

ShivamChoudhary17/Data-Science

Repository files navigation

Linear Regression :-

In statistics, linear regression is a linear approach to modelling the relationship between a dependent variable and one or more independent variables. It is both a statistical algorithm and a machine learning algorithm.

When there is a single input variable (x), the method is referred to as simple linear regression. When there are multiple input variables, literature from statistics often refers to the method as multiple linear regression.

Simple linear regression : 1 dependent variable (interval or ratio), 1 independent variable (interval or ratio or dichotomous).

Multiple linear regression : 1 dependent variable (interval or ratio) , 2+ independent variables (interval or ratio or dichotomous).

Logistic regression : 1 dependent variable (dichotomous), 2+ independent variable(s) (interval or ratio or dichotomous).

Note- Ordinary Least Squares and Gradient Descent are two common.

  • Four techniques to prepare a linear regression mode :-
  1. Simple Linear Regression

    image

With simple linear regression when we have a single input, we can use statistics to estimate the coefficients. This requires that you calculate statistical properties from the data such as means, standard deviations, correlations and covariance. All of the data must be available to traverse and calculate statistics.

  1. Ordinary Least Squares

    image

Used when we have more then one input. Given a regression line through the data we calculate the distance from each data point to the regression line, square it, and sum all of the squared errors together. This is the least squares seeks to minimize.

  1. Gradient Descent

    image A learning rate is used as a scale factor and the coefficients are updated in the direction towards minimizing the error. The process is repeated until a minimum sum squared error is achieved or no further improvement is possible.

    1. Regularization

      cd

      Regularization: Regularization is a common way of controlling or reducing the problem of overfitting in a flexible and tunable manner. Problem of overfitting is the result of model trying to capture noises with in the data.

      kx

About

Machine Learning, EDA, Feature Engg, PLot, Transformation of features

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published