Machine Learning By Andew Ng - Week 7
Large Margin Classification
Optimisation Objective
- 
    Alternative View Of Logistic Regression - 
        If y = 1, h(x) should be similar to 1, theta^T x » 0 
- 
        If y = 0, h(x) should be similar to 0, theta^T x » 1 
  - Graph plotting of function z
  
- 
        
- 
    Support Vector Machine - Modified hypothesis of Logistic Regression
    
Large Margin Intuition
- 
    Concept  
- 
    SVM Decision Boundary  - 
        SVM creates a decision boundary in a way that there is some space between the samples and decision boundary 
- 
        That space is called as margin 
  - 
        If C is too large the decision boundary will identical to magenta line 
- 
        If C is not to large then the decision boundary will be identical to black line 
  
- 
        
Mathematics Behind Large Margin Classification
- 
    Vector Inner Product - 
        u → euclidean length of vector u 
- p → length of projection of v onto u
  
- 
        
- 
    SVM Decision Boundary      
Kernels
Kernels I
- 
    Non Linear Decision Boundary  
- 
    Kernels  - 
        Similarity - Gaussian Kernel
  
- 
        Exmaple  
- 
        Concept  
 
- 
        
Kernels II
- 
    Choosing Landmarks  
- 
    SVM with Kernels    
- 
    SVM Parameters  
SVMs in Practice
Using An SVM
- 
    Overview  
- 
    Octave Implementation  
- 
    Other Kernels  
- 
    Multi Class Classification  
- 
    Logistic Regression VS Support Vector Machine 