Stemming Is The Process of Reducing Words To Their Base or Root Form (E.g., "Running"
Stemming Is The Process of Reducing Words To Their Base or Root Form (E.g., "Running"
1. What is NLP?
Answer: NLP (Natural Language Processing) is a subfield of artificial intelligence that focuses on the
interaction between computers and humans using natural language. It involves tasks like text
analysis, language generation, and understanding.
• Tokenization
• Sentiment Analysis
• Machine Translation
3. What is tokenization?
Answer: Tokenization is the process of splitting text into smaller units called tokens, which can be
words, phrases, or sentences.
4. What is stemming?
Answer: Stemming is the process of reducing words to their base or root form (e.g., "running" →
"run").
5. What is lemmatization?
Answer: Lemmatization reduces words to their base or dictionary form (lemma), considering the
context (e.g., "better" → "good").
Answer: Stop words are common words (e.g., "the", "is", "and") that are often removed to focus on
meaningful words in text analysis.
Answer: POS tagging assigns grammatical labels (e.g., noun, verb, adjective) to each word in a
sentence.
Answer: NER identifies and classifies entities in text into categories like names, dates, organizations,
etc.
9. What is a corpus?
Answer: A corpus is a large and structured collection of texts used for linguistic analysis and training
NLP models.
10. What is the difference between syntax and semantics?
Answer: Syntax refers to the structure of sentences, while semantics deals with the meaning of
words and sentences.
Answer: Word embedding is a dense vector representation of words that captures semantic
relationships (e.g., Word2Vec, GloVe).
Answer: Word2Vec is a neural network-based model that learns word embeddings by predicting
words in a context (CBOW) or predicting context from a word (Skip-gram).
Answer: GloVe (Global Vectors for Word Representation) is an unsupervised learning algorithm for
obtaining word embeddings by factorizing a word co-occurrence matrix.
Answer: A language model predicts the probability of a sequence of words, often used in text
generation and speech recognition.
Answer: An n-gram is a contiguous sequence of n items (words, characters) from a given text.
Answer: Sentiment analysis determines the emotional tone or opinion expressed in text (e.g.,
positive, negative, neutral).
Answer: Text normalization is the process of transforming text into a consistent format (e.g.,
lowercasing, removing punctuation).
Answer: Rule-based NLP uses handcrafted linguistic rules, while statistical NLP relies on machine
learning and data-driven approaches.
Answer: A confusion matrix is used to evaluate classification models by showing true positives, false
positives, true negatives, and false negatives.
22. What is precision and recall?
Answer: Precision measures the accuracy of positive predictions, while recall measures the fraction
of true positives correctly identified.
Answer: F1-score is the harmonic mean of precision and recall, providing a balance between the
two.
Answer: Overfitting occurs when a model performs well on training data but poorly on unseen data
due to excessive complexity.
Answer: Underfitting occurs when a model is too simple to capture patterns in the data, resulting in
poor performance on both training and test data.
27. What is the difference between supervised and unsupervised learning in NLP?
Answer: Supervised learning uses labeled data, while unsupervised learning works with unlabeled
data to find patterns.
Answer: A chatbot is an NLP application that simulates human conversation using text or voice.
Answer: Machine translation automatically translates text from one language to another (e.g.,
Google Translate).
Answer: Text summarization generates a concise summary of a longer text while retaining key
information.
Answer: Seq2Seq is a framework for tasks like machine translation, where an input sequence is
mapped to an output sequence using encoder-decoder architectures.
Answer: Attention mechanism allows models to focus on specific parts of the input sequence,
improving performance in tasks like translation.
33. What is Transformer architecture?
Answer: GPT (Generative Pre-trained Transformer) is a language model that uses autoregressive
transformers for text generation.
Answer: BERT is bidirectional and focuses on understanding context, while GPT is unidirectional and
focuses on text generation.
Answer: Transfer learning involves using pre-trained models (e.g., BERT, GPT) and fine-tuning them
for specific tasks.
Answer: Fine-tuning is the process of adapting a pre-trained model to a specific task by training it on
a smaller, task-specific dataset.
Answer: Zero-shot learning involves training a model to perform tasks it has never seen during
training, using generalizable knowledge.
Answer: Few-shot learning involves training a model with very few examples of a task.
Answer: A pre-trained language model is trained on a large corpus and can be fine-tuned for specific
NLP tasks.
Answer: Perplexity measures how well a language model predicts a sample, with lower values
indicating better performance.
Answer: Beam search is a decoding algorithm used in text generation to find the most likely
sequence of words.
Answer: Greedy search selects the most likely word at each step in text generation, without
considering future steps.
45. What is a dependency tree?
Answer: Coreference resolution identifies expressions that refer to the same entity in a text.
Answer: Semantic role labeling identifies the roles of words in a sentence (e.g., agent, patient).
Answer: LDA is a probabilistic model used for topic modeling, representing documents as mixtures
of topics.
Answer: Word sense disambiguation determines the correct meaning of a word based on context.
Answer: Self-supervised learning uses unlabeled data to create supervised tasks, such as predicting
masked words in BERT.
Answer: Masked language modeling involves predicting masked words in a sentence, used in models
like BERT.
Answer: Contrastive learning trains models to distinguish between similar and dissimilar pairs of
data points.
Answer: Multi-task learning trains a model on multiple related tasks simultaneously to improve
generalization.