IR Report
IR Report
Submitted By
Name: Atharva Litake
Roll no: 41143 Class: BE-1
2. PROBLEM DEFINITION
3. LEARNING OBJECTIVES
4. LEARNING OUTCOMES
5. ABSTRACT
8. CONCLUSION
1. TITLE: Text Summarization – Using Transformers and LLMs’
2. PROBLEM DEFINITION:
4. LEARNING OUTCOMES:
5. ABSTRACT:
1. Extractive Summarization:
- In extractive summarization, important sentences or phrases are directly
selected from the source text to create the summary.
- The process relies on identifying key sentences based on features such as term
frequency, sentence position, or using scoring models.
- Extractive methods tend to preserve the original structure of the text but may
lack coherence and flow in the summary.
2. Abstractive Summarization:
Abstractive summarization involves generating new sentences that capture the
essence of the input text, similar to how humans summarize. It requires deep
language understanding and relies on generating natural language summaries
using models such as transformers and LLMs. Abstractive summarization is
more complex and can generate coherent and fluent summaries but may
sometimes introduce factual inconsistencies.
Transformer Models in Summarization:
Transformer architectures like BART (Bidirectional and Auto-Regressive
Transformers) and its distilled variant DistilBART are commonly used for
abstractive summarization. These models consist of an encoder-decoder
structure. The encoder processes the input text, generating hidden
representations. The decoder generates the summary based on these
representations, often using attention mechanisms to focus on relevant parts of
the input.
DistilBART:
DistilBART is a compressed, faster version of BART, reducing the size and
computational cost while maintaining strong performance. It retains the core
transformer architecture but uses knowledge distillation techniques to create a
lighter model. DistilBART is suitable for tasks requiring both speed and
accuracy, making it a popular choice for large-scale text summarization.
Key Techniques:
Tokenization: Text is split into smaller units (tokens), which are then fed into the
transformer model.
Attention Mechanisms: The model weighs the importance of different words in
the input to generate contextually relevant summaries.
Pretraining and Fine-tuning: Pretrained models are further fine-tuned on specific
summarization datasets to specialize them for the summarization task.
Evaluation:
Common evaluation metrics include ROUGE (Recall-Oriented Understudy for
Gisting Evaluation) and BLEU (Bilingual Evaluation Understudy), which
compare the generated summary to reference summaries based on word overlaps
and n-gram matching.
7. GLIMPSE OF THE PROJECT & DEPLOYMENT:
8. CONCLUSION