Advances in AI: Module-1
Advances in AI: Module-1
Module-1
Distributional Semantics
Distributional semantics is an approach in
computational linguistics and Natural Language
Processing (NLP) that is based on the principle that the
meaning of a word can be inferred from the contexts in
which it appears. Distributional semantics leverages
large text corpora to model and quantify the meanings of
words based on their distributional properties—how
words co-occur with other words in text.
Distributional Semantics
1. Context and Co-occurrence
• Context: In distributional semantics, the context of a word
refers to the surrounding words or phrases in which it
appears. The context can be defined in several ways, such
as by using a fixed window of words around the target word
or by considering syntactic dependencies.
• Co-occurrence: The idea is that words that frequently
appear together in similar contexts tend to have related
meanings. For example, the words "king" and "queen" often
appear in similar contexts like "royalty," "throne," and
"palace," which suggests that they share a similar meaning.
Distributional Semantics
2. Vector Space Models (VSMs)
• Words are represented as vectors in a high-dimensional
space where each dimension corresponds to a specific
context feature (like the presence of another word in a given
window). The vector representation captures the meaning of
a word based on its distributional properties.
• Cosine Similarity: A common measure of similarity
between two word vectors is the cosine of the angle
between them. If two vectors are close in direction (high
cosine similarity), it suggests that the words they represent
are semantically similar.
Distributional Semantics
Word Embeddings
• Word2Vec: A neural network model that learns word embeddings by
predicting words based on their context (using either the Continuous Bag
of Words (CBOW) or Skip-gram model). Word2Vec produces dense,
continuous vector representations of words, where semantically similar
words are close in the vector space.
• GloVe (Global Vectors for Word Representation): This model
combines the advantages of global matrix factorization and local context
window methods. GloVe vectors are trained on aggregated global word-
word co-occurrence statistics from a corpus.
• FastText: An extension of Word2Vec that also considers subword
information (like character n-grams), allowing it to handle rare words or
misspellings more effectively.
• BERT (Bidirectional Encoder Representations from Transformers):
Unlike traditional word embeddings, BERT produces contextualized
Distributional Semantics
Applications of Distributional Semantics:
• Semantic Similarity: Distributional semantics is used
to compute the similarity between words, phrases, or
even sentences. For example, "dog" and "puppy" are
likely to have high similarity scores, while "dog" and
"car" will have a lower score.
• Word Sense Disambiguation: By comparing the
context in which a word appears, distributional
semantics can help disambiguate words with multiple
meanings. For example, "bank" in the context of "river"
vs. "money."
Distributional Semantics
• Information Retrieval: Search engines use
distributional semantics to improve the relevance of
search results by matching query terms with
semantically similar terms in documents.
• Text Classification: Distributional semantics is used to
represent text documents as vectors, which can then be
used for tasks like sentiment analysis, topic
classification, and spam detection.
• Machine Translation: Distributional models help in
aligning words and phrases between different
languages by mapping similar words to similar positions
in a shared vector space.
Frame Semantics
• Frame semantics plays a significant role in artificial
intelligence (AI), particularly in natural language
processing (NLP) and understanding. Incorporating
frame semantics into AI systems helps enhance their
ability to comprehend and generate human language in
a way that aligns more closely with human cognitive
processes.
Frame Semantics
Applications of Frame Semantics in AI