Linear regression
Linear regression is a linear model with constant slope unlike
logistic regression.

Predictive modeling is primarily concerned with minimizing the
error of a model or making the most accurate predictions possible, at the expense
of explainability. We borrow, reuse and steal algorithms from many different
fields, including statistics and use them towards these ends.
The representation of linear regression is an equation that
describes a line that best fits the relationship between the input variables
(x) and the output variables (y), by finding specific weightings for the
input variables called coefficients (B).
Linear regression has linear line in the graph.
For example:
y = B0 + B1 * x
We will predict y given the input x and the goal of the
linear regression learning algorithm is to find the values for the coefficients
B0 and B1.
Different techniques can be used to learn the linear regression
model from data, such as a linear algebra solution for ordinary least squares
and gradient descent optimization.
Linear regression has been around for more than 200 years and
has been extensively studied. Some good rules of thumb when using this
technique are to remove variables that are very similar (correlated) and to
remove noise from your data, if possible.
Comments
Post a Comment