Advances in AI: Module-1
Advances in AI: Module-1
Module-1
Co-Training
• Multiple Views: Co-training works under the assumption that the data can be
represented by multiple "views," where each view provides sufficient information
to make predictions. For example, in a text classification task, one view might be
based on the words in the text, and another view might be based on the
metadata (e.g., author, publication date).
• Iterative Process: The models are trained iteratively. Initially, they are trained on
a small labeled dataset. Then, each model makes predictions on the unlabeled
data. The most confident predictions are added to the labeled dataset, and the
models are retrained on this expanded dataset. This process is repeated until the
models converge.
• The effectiveness of MTL depends on the relatedness of the tasks. If the tasks are
too dissimilar, sharing representations might not be beneficial and could even
degrade performance.
• MTL can act as a form of regularization. By requiring the model to perform well
on multiple tasks, it discourages overfitting to a single task and promotes learning
more general features.
Multi-Task Learning
Applications
Task Imbalance: If one task is much harder or has more data than
others, it can dominate the learning process, leading to suboptimal
performance on the other tasks.
Negative Transfer: If the tasks are not sufficiently related, the model
might struggle, as knowledge from one task could negatively impact the
performance on another.
Coupled Semi-Supervised
Learning (CSSL)
Coupled Semi-Supervised Learning (CSSL) is an approach in machine
learning that combines elements of both supervised and unsupervised
learning to leverage the strengths of both paradigms. The core idea is
to use a small amount of labeled data in conjunction with a large
amount of unlabeled data to improve learning performance.
Coupled Semi-Supervised
Learning (CSSL)
• In traditional supervised learning, a model is trained on labeled data, where each
input comes with an associated label. In contrast, semi-supervised learning uses
both labeled and unlabeled data. The main goal is to improve model accuracy by
exploiting the structure in the unlabeled data to assist the learning process. This
is particularly useful when labeling data is expensive or time-consuming.
• Image and Video Analysis: CSSL can be used in image classification, where
different views might be different augmentations of the same image or features
extracted using different convolutional layers.
• Bioinformatics: CSSL is used in genomic data analysis, where one model might
focus on sequence data and another on structural data.
Coupled Semi-Supervised
Learning (CSSL)
Advantages
• Efficiency: It reduces the need for large amounts of labeled data, which can
be expensive and time-consuming to obtain.
Use Cases:
• Named Entity Recognition (NER): Identifying and classifying proper nouns (e.g., names of people,
places, organizations) within a text.
• Part-of-Speech Tagging: Determining the grammatical categories of each word (e.g., noun, verb,
adjective).
• Dependency Parsing: Analyzing the grammatical structure of a sentence to understand how
words relate to each other.
• Question Answering: Finding precise answers to questions within a specific passage of text.
Macro reading vs Micro reading
Techniques
• Ambiguity and Polysemy: Words and phrases can have multiple meanings
depending on the context, making it challenging for Open IE systems to
accurately extract the correct relations.
• Complex Sentences: Sentences with complex structures, such as nested clauses
or multiple entities and relations, can be difficult for Open IE systems to parse
correctly.
• Lack of Schema: While the lack of a predefined schema provides flexibility, it can
also lead to inconsistencies in the extracted information, as the same relation
might be represented in multiple ways.
QUESTIONS
• Name some popular Open IE.
• What challenges arise when applying coupled semi-supervised
learning to domains with highly imbalanced datasets, and how can
these challenges be mitigated to ensure robust model training?
• In what scenarios is coupled semi-supervised learning most effective,
and how does it compare to other learning paradigms like co-training
or multi-task learning in terms of data efficiency and generalization?