Linear Classifiers are one of the most commonly used classifiers and Logistic Regression is one of the most commonly used linear classifiers. The concepts we are going to learn here will actually extend a lot of other classification methods beyond linear classifiers. We’re going to learn the fundamental concepts, but also the underlying algorithms that let us optimize the parameters of this model to fit your training data.
Suppose, we want to make a spam filter which will filter spams from our email list automatically. So, for every email, we’re going to feed it to a classifier and the classifier will say if this is a spam or maybe we want to visit a restaurant for a specific food e.g. sushi but before visiting the restaurant we want to make sure that the sushi there is good. So, what we want to do is to check the review of others visiting this restaurant about sushi and see if it is positive. It is actually very difficult to manually check all the reviews. Instead, what we can do is to feed these reviews to a classifier and it is going to say, is this a positive sentiment or negative sentiment and then watching the number of positive sentiment and the negative sentiment we will decide whether the sushi there is good to go or not.
Linear Classifiers Intuition
A linear classifier will take some quantity x as input which in our case will be emails or reviews and is going to make a prediction ŷ that says is this a positive statement which means is this a no spam or a positive review in which case ŷ = +1 or is this a negative statement which implies it is a spam or a negative review in which case ŷ = -1. We will use restaurant review classifier for the rest fo the tutorial as our use case. A linear classifier does a little bit more associating every word for weight or coefficient which says how positively or negatively influential this word is for a review. For example,
Here good and delicious have a coefficient of 1.0 and 1.5 respectively. Wonderful is very positive and has a coefficient of 2.7. On the negative side, bad and terrific might have a coefficient of -1 and -2.1 respectively. But awful is just awful, so -3.3. There are also some words that are not that relevant to the sentiment of the review, might have 0 coefficient.
Now how will we use these coefficient’s to make a prediction of whether a sentence is positive or negative? Let’s take a review, for example,
here good, delicious and great have coefficients greater than 0 and rest of the terms having coefficients greater than 0 as they are not relevant. After calculating we get a positive score which implies the sentiment in the sentence is positive and ŷ = +1. This is called a linear classifier because the output is the weighted sum of the inputs.
So more generally for a simple linear classifier, we are going to take a review and the coefficient associated with each word and feed that in our classifier and calculate the score for that input. If the score is greater than 0 the prediction is positive and ŷ = +1 and if the score is less than 0 we say the prediction is ŷ = -1. Now, what we need to do is to train the coefficients or weights of these linear classifiers from data.
So given some input training data that includes sentences of reviews labeled with either +1 or -1. We’re going to split those into some training set and some validation set. Then we’re going to feed that training set to some learning algorithm which is going to learn the weights associated with each word. And then after we learn this classifier, we’re going back and evaluate its accuracy on that validation set. Now, how do we learn this classifier from data?
The decision boundary is a boundary between positive and negative predictions. Let’s say that we have taken our data and trained our linear classifier and every word has zero weight except for two of them. Awesome has weight 1.0 and awful has weight -1.5. That means that the score of any sentence is 1.0 times the number of the word awesome minus 1.5 times the number of times the word awful shows up. We can plot that into a graph which depends on every sentence the number of awesome and the number of awful. So for example, for a sentence,
We’re going to plot that into space where we’re going to have two awesome and one awful. So it gets plotted in the (2,1) point. And then for every sentence that we might have in our training data set say, three awful and one awesome, three awesome and no awful and so on we will have different points and if we plot the dataset we find
The classifier that we’ve trained with the coefficients 1.0 and -1.5 will have a decision boundary that corresponds to a line plotted above, where 1.0 times awesome minus 1.5 times the number of awful is equal to zero. Everything below that line has a score greater than zero and any points above that line are going to have scored less than zero. So there’s that line, everything below the line is positive, everything above the line’s negative a linear decision boundary. That is what makes it a linear classifier.
Linear Classifiers Model
in our previous example that we had with just two features with no zero coefficients, awesome and awful and we have calculated the score as
where w₀ was 0 in our example. Now suppose that we had a third feature with no zero coefficient. Let’s say that the word great. So the score function then will be
So, the general form of our linear classifier model will be :
Theses notations that we’ve used so far are without features associated with it. So we’re going to have these functions hᵢ where i is 1 through D. They’ve defined some features we might extract from the data and we are going to encode the constant function that is h₀. So the more generically our model will be looked like:
So this is our generic linear classifier model with multiple features.
Linear classifiers are a pretty abstract concept. Logistic regression is a specific case of that, where we use what’s called the logistic function to squeeze minus infinity to plus infinity into the interval (0, 1) so we can predict probabilities for every class. Next, we will discuss Logistic Regression.
Originally published at Gadictos.