- Regression
A prediction comes out within a range of any continuous value such as weight, age, speed.
- Linear Regression
It is used for finding linear relationship between dependent and one or more predictors.
| x------>y affect x : affect to y, predictor or independent variable y : subordinated by x, response or dependent variable |
1. Simple Linear Regression
Presume one x.
| y=Wx+b W : Weight, gradient b : Bias, intercept |
2. Multiple Linear Regression
Presume more than one x.
| y=W1x1+W2x2+...Wnxn+b |

3. Cost function(Loss function, Objective function)
Error=real value-predict value![]() |
| Minimize the error. W,b→minimize cost(W,b) 0 is the best! ![]() |
4. MSE
Mean Squared Error.
![]() |
5. Optimizer(Gradient Descent)
The process to find W,b for minimizing cost function.


'Machine Learning' 카테고리의 다른 글
| Entropy, Cross-Entropy (0) | 2021.03.31 |
|---|---|
| Support Vector Machine, Margin, Kernel, Regularization, Gamma (0) | 2021.03.30 |
| Softmax (0) | 2021.03.24 |
| Logistic Regression, Sigmoid function (0) | 2021.03.17 |
| Scalar vs Vector vs Matrix vs Tensor (0) | 2021.03.17 |


