Interview Preparation of SVM.
Most Important Interview topics in SVM..
- SVM
- I have written all the basic and maths intuition of Linear SVM in Paper so that it’s easy to comprehend:
- This is All About Linear SVM. For non linear SVM we use Kernal Trick that are available in scikit learn.
- 1. What Are the Basic Assumption?
- 2. Advantages
- 3. Disadvantages
- 4. Whether Feature Scaling is required?
- 5. Impact of Missing Values?
- 6. Impact of outliers?
- 7. Types of Problems it can solve(Supervised)
- 8. Overfitting And Underfitting
- 9. Different Problem statement you can solve using Naive Baye’s
- 10. Practical Implementation
- 11. Performance Metrics
SVM
Theoretical Understanding:
I have written all the basic and maths intuition of Linear SVM in Paper so that it’s easy to comprehend:
This is All About Linear SVM. For non linear SVM we use Kernal Trick that are available in scikit learn.
1. What Are the Basic Assumption?
There are no such assumptions
2. Advantages
- SVM is more effective in high dimensional spaces.
- SVM is relatively memory efficient.
- SVM’s are very good when we have no idea on the data.
- Works well with even unstructured and semi structured data like text, Images and trees.
- The kernel trick is real strength of SVM. With an appropriate kernel function, we can solve any complex problem.
- SVM models have generalization in practice, the risk of over-fitting is less in SVM.
3. Disadvantages
- More Training Time is required for larger dataset
- It is difficult to choose a good kernel function
- The SVM hyper parameters are Cost -C and gamma. It is not that easy to fine-tune these hyper-parameters. It is hard to visualize their impact
4. Whether Feature Scaling is required?
Yes
5. Impact of Missing Values?
Although SVMs are an attractive option when constructing a classifier, SVMs do not easily accommodate missing covariate information. Similar to other prediction and classification methods, in-attention to missing data when constructing an SVM can impact the accuracy and utility of the resulting classifier.
6. Impact of outliers?
It is usually sensitive to outliers Link To Paper
7. Types of Problems it can solve(Supervised)
- Classification
- Regression
8. Overfitting And Underfitting
In SVM, to avoid overfitting, we choose a Soft Margin, instead of a Hard one i.e. we let some data points enter our margin intentionally (but we still penalize it) so that our classifier don’t overfit on our training sample
Detail about SVM in Scikit-Learn
9. Different Problem statement you can solve using Naive Baye’s
- We can use SVM with every ANN usecases
- Intrusion Detection
- Handwriting Recognition
10. Practical Implementation
Link to Practical Implementation for SVC
Link to Practical Implementation for SVR
11. Performance Metrics
Classification
- Confusion Matrix
- Precision,Recall, F1 score
Regression
- R2,Adjusted R2
- MSE,RMSE,MAE