Multiclass Classification: One-vs-all
Classify more than just 0 and 1. Examples:- Email foldering/tagging: Work (y=1), Friends (2), Family (3), Hobby (4)
- Medical diagrams: Not ill (1), Cold (2), Flu (3)
- Weather: Sunny (1), Cloudy (2), Rain (3), Snow (4)
Summary by Filip:
- Logistic regression is a classification algorithm and not a regression algorithm, although "regression" is part of the name
- Hypothesis equals sigmoid/logistic function of theta_transpose * x
- The derivative of J(theta) is the same as for linear regression, i.e. 1/m * sum((h(x) - y)*x). However, the hypothesis is different. While the hypothesis in linear regression is h(x) = theta_transpose * X , the hypothesis in logistic regression is 1/(1+e^(theta_transpose * x))
- Function z does not have to be linear; it can be everything, e.g. circle
- When dealing with multiclass logistic regression, if we follow the one vs. all approach, we can simplify the problem to simple binary classification problem. So we estimate P(y=i|x;theta) for i={1,..n} for each class and take the max
Resources:
Comments
Post a Comment