AI4youngster - 6 - Topic NLP
AI4youngster - 6 - Topic NLP
2. Overview of NLP
3. Method in NLP
4. Trend in NLP
5. Conclusion
1. Some achievements of NLP
NLP in Industry
Communication With Machines
Virtual Assistant
Understanding (NLU)
Generation
(NLG)
Natural language processing and computational linguistics
● Natural language processing (NLP) develops methods for solving practical problems
involving language:
○ Automatic speech recognition
○ Machine Translation
○ Sentiment Analysis
○ Information extraction from documents
● Computational linguistics (CL) focused on using technology to support/implement
linguistics:
○ How do we understand language?
○ How do we produce language?
○ How do we learn language?
Level Of Linguistic Knowledge
Morphology
● Syntax studies the ways words combine to form phrases and sentences
● Syntactic parsing helps identify who did what to whom, a key step in understanding a
sentence
Semantics and pragmatics
● Ambiguity
● Sparsity
● Abstractly, most NLP applications can be viewed as prediction problems
Should be able to solve them with Machine Learning
● The label set is often the set of all possible sentences
○ Infinite (or at least astronomically large)
● Training data for supervised learning is often not available
Unsupervised/semi-supervised techniques for training from available data
● Algorithmic challenges
○ Vocabulary can be large (e.g., 50K words)
○ Data sets are often large (GB or TB)
Ambiguity
● Up until 2014 most state-of-the-art NLP systems were based on feature engineering +
shallow machine learning models (e.g., SVMs, CRFs)
● Designing the features of a winning NLP system requires a lot of domain-specific
knowledge.
● Deep Learning systems on the other hand rely on neural networks to automatically
learn good representations.
Feature Engineering and Deep Learning
● Remove information no longer needed from the context, and add information likely to
be needed later
● Do this by:
○ Adding an explicit context layer to the architecture
○ This layer controls the flow of information into and out of network layers using specialized neural units called gates
Encoder-decoder Framework
● Seq2Seq (Transformer)
● Transfer Learning. Fine-tuning a pre-trained language model (LM) has become the de
facto standard for doing transfer learning in natural language processing
● Combining Supervised & Unsupervised Methods
● Self supervised learning
● Reinforcement Learning
● NLP model with interpretability
● Combine Feature Engineering and Knowledge base
Trends in NLP for applications
http://anthology.aclweb.org/
Conclusion