Mixed ML Q&A - Set 2
20 Core Questions
Interview Prep
Mixed Machine Learning Concepts: Q&A (Set 2)
Short mixed-topic questions on interpretability, fairness, experimentation and ML system design.
Fairness
Experimentation
Explainability
Systems
1
What is model interpretability and why does it matter?
⚡ Beginner
Answer: Interpretability is the ability to understand and explain why a model made a prediction, which is critical for trust, debugging and regulation.
2
Name some techniques for explaining ML model predictions.
📊 Intermediate
Answer: Techniques include feature importance, partial dependence plots, SHAP, LIME and counterfactual explanations.
3
What is SHAP in simple terms?
🔥 Advanced
Answer: SHAP uses ideas from Shapley values in game theory to attribute how much each feature contributed to an individual prediction.
4
What does fairness mean in ML systems?
🔥 Advanced
Answer: Fairness roughly means similar individuals or groups are treated similarly; metrics often focus on error or decision parity across protected groups.
5
Name two common fairness metrics.
🔥 Advanced
Answer: Examples: demographic parity, equal opportunity, equalized odds.
6
Why can optimizing purely for accuracy be risky in production ML?
📊 Intermediate
Answer: Accuracy may hide minority errors, fairness issues, calibration problems or misalignment with business outcomes.
7
What is an A/B test in the context of ML systems?
⚡ Beginner
Answer: An A/B test compares two versions of a system (A vs B) on real users to see which performs better on chosen metrics.
8
What is conceptually different between offline validation and online A/B testing?
🔥 Advanced
Answer: Offline validation uses historical data and proxy metrics; A/B tests measure live impact on business KPIs with real users.
9
What is MLOps and why is it important?
📊 Intermediate
Answer: MLOps applies DevOps principles to ML, focusing on reliable training, deployment, monitoring and iteration of models in production.
10
What is a feature store and when would you use one?
🔥 Advanced
Answer: A feature store centralizes feature definitions, storage and serving for both training and inference, improving consistency and reuse.
11
How do you handle model versioning in an ML project?
📊 Intermediate
Answer: Use tools or conventions to track code, data, hyperparameters and artifacts per version (e.g., Git + MLflow/DVC).
12
Why is data quality often more critical than model choice?
⚡ Beginner
Answer: No model can fix systematic errors, label noise, or missing coverage; clean, representative data sets the ceiling on performance.
13
What is active learning in ML?
🔥 Advanced
Answer: Active learning lets the model choose which unlabeled examples to label next, aiming to improve accuracy with fewer labels.
14
What is transfer learning and when is it helpful?
📊 Intermediate
Answer: Transfer learning reuses a model trained on a related task or large dataset, then fine-tunes it on a smaller, specific dataset.
15
Why is feature importance not the same as causality?
🔥 Advanced
Answer: High importance shows a predictive association, not necessarily that changing the feature will change the outcome.
16
What are some common sources of label noise in real datasets?
📊 Intermediate
Answer: Label noise comes from manual annotation errors, ambiguous cases, outdated labels, or automated heuristics.
17
How do you decide whether to invest in a more complex model or better data?
🔥 Advanced
Answer: If simple models already hit a data-quality ceiling, better data is key; if metrics improve with complexity, model work may pay off, but always compare to baselines.
18
What is a data pipeline in ML systems?
⚡ Beginner
Answer: The data pipeline covers extracting, cleaning, transforming and delivering data consistently to both training and inference environments.
19
Why is it important to align ML metrics with business goals?
⚡ Beginner
Answer: A model can optimize a technical metric but still hurt revenue, user experience or risk if metrics aren’t aligned with real objectives.
20
What is the main takeaway from this advanced mixed set?
⚡ Beginner
Answer: Modern ML practice combines solid modeling with careful design of experiments, fairness, monitoring and system engineering.
Quick Recap: Mixed ML Concepts 2
Thinking beyond algorithms—about users, systems and long-term behavior—is what makes ML solutions reliable and responsible in the real world.