LLM1
LLM1
calls.
Here’s how I’d approach it with current technology:
3. Conversation Summarization
Model Choice: Use a summarization model like PEGASUS, BART, or GPT-4-turbo.
For summarization → FLAN-T5 or Gemini Pro. FLAN-T5 (lighter, more efficient for cost-sensitive applications). Gemini Pro (better
performance for longer conversations).
Pipeline:
Transcript Cleaning: Remove filler words, false starts, etc.
Chunking for Long Calls: Use a sliding window approach if the call is longer than the token limit.
This would be kind of a mess b/c you have to chunk, then do an overlapping chunk, summarize each one, and concatenate, then get the
model to take that concatenation and summarize THAT to make it make sense. Could end up redundant or missing the overarching
meaning.
You can also add as you go like start with the first chunk’s summary and feed that into the model as part of the context for the second
chunk’s summary request, then build up over time.
Summarization Strategy: Use extractive summarization for factual accuracy (e.g., keyword extraction with embeddings) and abstractive
summarization for readability.
Compliance & Sentiment Analysis: Add an extra step for detecting tone, key customer concerns, and compliance risks.
Evaluation Methods
Chatbot:
Response accuracy – how many of the classifications did it get right
Intent recognition accuracy – how many intents did it get right
Generated response or summary – BLEU or ROUGE to compare to human responses
Ask users in survey, check if they engage again with bot
Reinforcement Learning (RL) is related to NLP in several ways, particularly in areas where models need to generate or refine outputs based on
feedback rather than direct supervision. Some key applications include:
PPO's objective function balances maximizing rewards while keeping policy updates stable:
This function optimizes rewards while avoiding drastic changes to the policy.
Graph-based NLP techniques represent linguistic structures as graphs and apply graph algorithms or neural networks to analyze text. These
methods are useful for tasks where relationships between words, sentences, or documents matter.
🔹 Example: In keyword extraction, words that co-occur frequently form a graph. The most "central" words (highly connected) are
extracted as keywords.
Models like Graph Convolutional Networks (GCN) or Graph Attention Networks (GAT) process NLP graphs.
Used in relation extraction, document classification, and entity linking.
Example: A GNN-based citation network classifies research papers by linking them in a graph.
C. Knowledge Graphs