SVM Q&A 20 Core Questions
Interview Prep

Support Vector Machines: Interview Q&A

Short questions and answers on SVMs: margins, kernels, hyperparameters and when to prefer them.

Margin Kernel Trick C & Gamma Support Vectors
1 What is the main idea behind SVMs for classification? âš¡ Beginner
Answer: SVMs aim to find a hyperplane that maximizes the margin between classes, i.e., the distance to the closest points.
2 What are support vectors? âš¡ Beginner
Answer: Support vectors are the training points closest to the decision boundary; they determine the position of the hyperplane.
3 What is the margin in SVMs? âš¡ Beginner
Answer: The margin is the distance between the separating hyperplane and the closest data points from each class.
4 What is the soft‑margin SVM? 📊 Intermediate
Answer: Soft‑margin SVM allows some misclassifications by introducing slack variables, balancing margin size and classification errors.
5 What does the C parameter control in SVMs? 📊 Intermediate
Answer: C controls the trade‑off between maximizing margin and minimizing classification error; large C focuses on correctly classifying training points, small C allows more margin violations.
6 What is the kernel trick in SVMs? 🔥 Advanced
Answer: The kernel trick lets SVMs implicitly operate in a higher‑dimensional feature space without computing coordinates explicitly, using kernel functions.
7 Name some common SVM kernels. âš¡ Beginner
Answer: Common kernels include linear, polynomial, RBF (Gaussian) and sigmoid.
8 What does the gamma parameter mean for RBF kernels? 🔥 Advanced
Answer: Gamma controls the influence of individual training examples; high gamma means each point has a small radius of influence, leading to complex boundaries.
9 How do C and gamma affect bias‑variance in RBF SVMs? 🔥 Advanced
Answer: High C and high gamma yield low bias, high variance models; low C and low gamma give high bias, low variance models.
10 Do SVMs require feature scaling? 📊 Intermediate
Answer: Yes, SVMs are sensitive to feature scales, especially with RBF or polynomial kernels; scaling usually improves performance and convergence.
11 Are SVMs good for very large datasets? 🔥 Advanced
Answer: Classic SVMs can be slow and memory‑intensive for very large datasets; linear SVMs or approximate methods are often preferred.
12 Can SVMs be used for regression? 📊 Intermediate
Answer: Yes, SVR (Support Vector Regression) is the regression counterpart, using an epsilon‑insensitive loss.
13 How do you extend SVMs to multi‑class problems? 📊 Intermediate
Answer: Common strategies are one‑vs‑rest and one‑vs‑one, training multiple binary SVMs and combining their outputs.
14 Do SVMs output calibrated probabilities by default? 🔥 Advanced
Answer: No, SVM scores are not probabilities; techniques like Platt scaling or isotonic regression are used for calibration.
15 When is a linear SVM a good choice? âš¡ Beginner
Answer: Linear SVMs work well for high‑dimensional, approximately linearly separable data such as text classification.
16 What are some advantages of SVMs? âš¡ Beginner
Answer: SVMs can achieve strong performance, work in high dimensions, and handle non‑linear boundaries with kernels.
17 What are some disadvantages of SVMs? âš¡ Beginner
Answer: They can be slow on large datasets, sensitive to hyperparameters and less interpretable than linear models or trees.
18 How do you tune C and gamma in practice? 📊 Intermediate
Answer: Typically via grid search or randomized search with cross‑validation, exploring logarithmic ranges of C and gamma.
19 Give a real‑world use case where SVMs have been successful. ⚡ Beginner
Answer: SVMs have been widely used in text categorization, image classification and bioinformatics (e.g., gene expression analysis).
20 What is the key message to remember about SVMs? âš¡ Beginner
Answer: SVMs are powerful margin‑based classifiers; if you understand margins, kernels and the roles of C and gamma, you can tune and use them effectively.

Quick Recap: SVMs

SVMs shine when data is high‑dimensional and you can afford careful tuning; think in terms of margin plus kernel to explain and reason about them in interviews.