Mixed NLP Q&A – Set 2

Mixed NLP questions – core concepts set 2

20 more mixed questions and answers that test your understanding of sequence labeling, NER, QA, summarization, attention, retrieval and evaluation across the NLP stack.

1

What is sequence labeling and which NLP tasks fit this formulation?

Answer: Sequence labeling assigns a label to each token in a sequence; tasks include POS tagging, chunking, NER, BIO tagging for entities and token-level classification in many structured prediction problems.

2

Why are CRF layers often added on top of BiLSTM or transformer encoders for NER?

Answer: CRFs model dependencies between labels (e.g. valid BIO transitions), enforcing global consistency across the sequence and reducing inconsistent tag sequences that a per-token classifier might produce independently.

3

What is the difference between extractive and abstractive summarization at a high level?

Answer: Extractive methods select sentences or phrases verbatim from the source, while abstractive methods generate new text that may paraphrase or compress the content, similar to how humans write summaries.

4

How does multi-hop question answering differ from simple factoid QA?

Answer: Multi-hop QA requires reasoning across multiple pieces of evidence, often from different sentences or documents, whereas factoid QA can be answered from a single local span that directly contains the fact.

5

What is attention masking and where is it used?

Answer: Attention masks prevent attending to certain positions, such as padding tokens or future tokens in causal language models, ensuring correct semantics and avoiding leakage of information the model should not see.

6

Why is calibration of model confidence important in deployed NLP systems?

Answer: Calibrated probabilities allow systems to know when to abstain, defer to humans or request clarification, reducing harmful mistakes and enabling better risk management in high-stakes applications like medicine or finance.

7

What role does a retriever play in retrieval-augmented generation?

Answer: The retriever searches external knowledge sources (like vector indices or BM25 search) for relevant documents based on a query, and the generator then conditions on this retrieved context to produce more accurate, grounded responses.

8

Why is it useful to separate retrieval and generation components in NLP systems?

Answer: Separation allows independent optimization, easier updates of knowledge (by changing the index instead of retraining the LM) and modular design where different retrievers or generators can be swapped for various tasks and constraints.

9

What is label smoothing and why is it used in sequence models?

Answer: Label smoothing replaces one-hot target distributions with slightly softened ones, preventing the model from becoming overconfident and often improving generalization, especially in encoder–decoder generation tasks like MT.

10

How can evaluation metrics encourage undesirable behavior in generative models?

Answer: If metrics reward surface overlap or length characteristics too strongly, models may learn to game them by copying source text, being overly verbose or repeating common phrases, rather than optimizing for user satisfaction or correctness.

11

What is catastrophic forgetting in the context of fine-tuning language models?

Answer: Catastrophic forgetting occurs when fine-tuning on a small dataset overwrites useful general knowledge learned during pretraining, degrading performance on previous capabilities, which can be mitigated via careful learning rates and regularization.

12

Why might you use adapter layers instead of full fine-tuning?

Answer: Adapters add small trainable modules within a frozen backbone, significantly reducing the number of parameters to update, easing storage of multiple task adaptations and mitigating forgetting across tasks sharing a base model.

13

What is the difference between extractive QA and open-domain QA?

Answer: Extractive QA answers questions from a specified context passage by selecting spans, while open-domain QA must first locate relevant documents from a large corpus or web, then extract or generate the answer from those sources.

14

How can QA-based evaluation be used to assess summarization quality?

Answer: QA-based evaluation creates questions from the source document and checks whether answers can be recovered from the summary, indirectly measuring how much critical information the summary preserves compared to the original text.

15

What are hallucinations in retrieval-augmented systems and why can they still occur?

Answer: Even with retrieval, generators may ignore or misinterpret documents and invent details; retrieval reduces but does not eliminate hallucinations, so careful prompt design, grounding checks and output validation remain necessary.

16

Why are system and developer messages important in chat-based LLM APIs?

Answer: They provide high-priority instructions and constraints that shape the assistant’s behavior across many turns, enabling consistent persona, policy enforcement and separation between user content and system-level rules.

17

What is data leakage and how can it affect NLP model evaluation?

Answer: Data leakage occurs when test information unintentionally appears in training or prompts; it leads to over-optimistic performance estimates because the model effectively memorizes answers rather than generalizing from unseen data.

18

Why should evaluation sets be representative and diverse?

Answer: Representative, diverse evaluation sets uncover weaknesses across topics, styles and user groups, avoiding misleading conclusions that a system is strong when it only performs well on narrow or biased test slices.

19

How does curriculum learning sometimes help in training NLP models?

Answer: Curriculum learning presents easier examples first and gradually increases difficulty, which can stabilize optimization and help models learn better representations before tackling more complex inputs or tasks.

20

What mindset is helpful when preparing for mixed-topic NLP interviews?

Answer: Focus on understanding core principles rather than memorizing details, practice explaining trade-offs and design decisions clearly and be ready to connect concepts across modeling, data, evaluation and deployment concerns.

🔍 Mixed NLP concepts covered – Set 2

This page reinforces advanced NLP concepts: structured prediction, QA and summarization, attention and RAG patterns, calibration, adapters, evaluation design and deployment issues important for real interview discussions.

Sequence labeling & CRFs
QA & multi-hop reasoning
Attention masks & RAG
Fine-tuning & adapters
Metric pitfalls & leakage
Interview preparation mindset