Mixed ML Q&A - Set 2 20 Core Questions
Interview Prep

Mixed Machine Learning Concepts: Q&A (Set 2)

Short mixed-topic questions on interpretability, fairness, experimentation and ML system design.

Fairness Experimentation Explainability Systems
1 What is model interpretability and why does it matter? ⚡ Beginner
Answer: Interpretability is the ability to understand and explain why a model made a prediction, which is critical for trust, debugging and regulation.
2 Name some techniques for explaining ML model predictions. 📊 Intermediate
Answer: Techniques include feature importance, partial dependence plots, SHAP, LIME and counterfactual explanations.
3 What is SHAP in simple terms? 🔥 Advanced
Answer: SHAP uses ideas from Shapley values in game theory to attribute how much each feature contributed to an individual prediction.
4 What does fairness mean in ML systems? 🔥 Advanced
Answer: Fairness roughly means similar individuals or groups are treated similarly; metrics often focus on error or decision parity across protected groups.
5 Name two common fairness metrics. 🔥 Advanced
Answer: Examples: demographic parity, equal opportunity, equalized odds.
6 Why can optimizing purely for accuracy be risky in production ML? 📊 Intermediate
Answer: Accuracy may hide minority errors, fairness issues, calibration problems or misalignment with business outcomes.
7 What is an A/B test in the context of ML systems? ⚡ Beginner
Answer: An A/B test compares two versions of a system (A vs B) on real users to see which performs better on chosen metrics.
8 What is conceptually different between offline validation and online A/B testing? 🔥 Advanced
Answer: Offline validation uses historical data and proxy metrics; A/B tests measure live impact on business KPIs with real users.
9 What is MLOps and why is it important? 📊 Intermediate
Answer: MLOps applies DevOps principles to ML, focusing on reliable training, deployment, monitoring and iteration of models in production.
10 What is a feature store and when would you use one? 🔥 Advanced
Answer: A feature store centralizes feature definitions, storage and serving for both training and inference, improving consistency and reuse.
11 How do you handle model versioning in an ML project? 📊 Intermediate
Answer: Use tools or conventions to track code, data, hyperparameters and artifacts per version (e.g., Git + MLflow/DVC).
12 Why is data quality often more critical than model choice? ⚡ Beginner
Answer: No model can fix systematic errors, label noise, or missing coverage; clean, representative data sets the ceiling on performance.
13 What is active learning in ML? 🔥 Advanced
Answer: Active learning lets the model choose which unlabeled examples to label next, aiming to improve accuracy with fewer labels.
14 What is transfer learning and when is it helpful? 📊 Intermediate
Answer: Transfer learning reuses a model trained on a related task or large dataset, then fine-tunes it on a smaller, specific dataset.
15 Why is feature importance not the same as causality? 🔥 Advanced
Answer: High importance shows a predictive association, not necessarily that changing the feature will change the outcome.
16 What are some common sources of label noise in real datasets? 📊 Intermediate
Answer: Label noise comes from manual annotation errors, ambiguous cases, outdated labels, or automated heuristics.
17 How do you decide whether to invest in a more complex model or better data? 🔥 Advanced
Answer: If simple models already hit a data-quality ceiling, better data is key; if metrics improve with complexity, model work may pay off, but always compare to baselines.
18 What is a data pipeline in ML systems? ⚡ Beginner
Answer: The data pipeline covers extracting, cleaning, transforming and delivering data consistently to both training and inference environments.
19 Why is it important to align ML metrics with business goals? ⚡ Beginner
Answer: A model can optimize a technical metric but still hurt revenue, user experience or risk if metrics aren’t aligned with real objectives.
20 What is the main takeaway from this advanced mixed set? ⚡ Beginner
Answer: Modern ML practice combines solid modeling with careful design of experiments, fairness, monitoring and system engineering.

Quick Recap: Mixed ML Concepts 2

Thinking beyond algorithms—about users, systems and long-term behavior—is what makes ML solutions reliable and responsible in the real world.