Answer
Prerequisites
- Hyperplane Equation: From part (b), the decision boundary is .
- Mahalanobis Distance: The squared Mahalanobis distance between two vectors and with respect to covariance matrix is defined as .
- Vector Transpose Properties: For a scalar value resulting from a vector product , its transpose is equal to itself: . If is symmetric (like ), then .
Step-by-Step Derivation
-
Target Form: We want to rewrite the hyperplane equation into the form . Expanding the target form gives: Comparing this with , we can see that we must satisfy the condition:
-
Substitute Knowns: From part (b), we have: We are given the proposed expression for :
-
Evaluate : Let's calculate using the given definitions and show it equals . Since is symmetric, .
-
Distribute the Terms: Multiply into the brackets:
-
Simplify the Expression:
- First term: Notice that is a scalar. Its transpose is . Since a scalar equals its transpose, we can rewrite the first term as .
- Second term: By definition, the numerator is exactly the squared Mahalanobis distance . Therefore, the fraction cancels out to 1.
Substituting these simplifications back:
-
Conclusion: We have shown that . Therefore, the equation is perfectly equivalent to , which is .
Interpretation
- Interpretation of : The vector is the normal vector to the decision hyperplane. It determines the orientation (tilt) of the boundary. It points in the general direction from the mean of class to the mean of class , but is skewed by the inverse covariance matrix to account for the shape and spread of the data distribution.
- Interpretation of : The point is a specific point that lies exactly on the decision hyperplane (since ). It acts as an anchor point or origin for the boundary.
- Effect of the priors on :
The formula for consists of two parts: a midpoint and a shift term.
- If the classes are equally probable (), then . The shift term disappears, and is exactly halfway between the two class means.
- If class is more probable (), then . The shift term subtracts a vector pointing from to . This moves the anchor point away from and towards . Geometrically, this shifts the entire decision boundary towards the less probable class , thereby expanding the decision region assigned to the more probable class .