0% found this document useful (0 votes)
63 views9 pages

NLP Class10 PDF

Natural Language Processing (NLP) is a sub-field of AI that enables computers to understand human languages through various processes such as audio capture, text conversion, and response generation. Applications of NLP include chatbots, voice assistants, language translation, and sentiment analysis, with chatbots simulating human conversation using AI. Key steps in NLP text analysis include sentence segmentation, tokenization, stemming, and lemmatization, highlighting the differences between human and computer language interpretation.

Uploaded by

aryasreegopi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
63 views9 pages

NLP Class10 PDF

Natural Language Processing (NLP) is a sub-field of AI that enables computers to understand human languages through various processes such as audio capture, text conversion, and response generation. Applications of NLP include chatbots, voice assistants, language translation, and sentiment analysis, with chatbots simulating human conversation using AI. Key steps in NLP text analysis include sentence segmentation, tokenization, stemming, and lemmatization, highlighting the differences between human and computer language interpretation.

Uploaded by

aryasreegopi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 9

NATURAL LANGUAGE PROCESSING

NLP, is the sub-field of AI that is focused on enabling computers


to understand and process human languages.
In NLP, we can break down the process of understanding English for a model into a number of
small pieces.
A usual interaction between machines and humans using Natural Language Processing
could go as follows:
Humans talk to the computer
The computer captures the audio
There is an audio to text conversion
Text data is processed Data is converted to audio
The computer plays the audio file and responds to humans

Applications of Natural Language Processing


Chatbots
Chatbots are a form of artificial intelligence that is programmed to interact
with humans in such a way that they sound like humans themselves.

Chatbots are created using Natural Language Processing and Machine


Learning.
Chatbots work in two simple steps.
First, they identify the meaning of the question asked and collect all the data from the user that
may be required to answer the question.
Then they answer the question appropriately.

2. Autocomplete in Search Engines


3.Voice Assistants
4.Language Translator
5.Sentiment Analysis
6.Grammar Checkers
7.Email Classification and Filtering
8.Text Summarization

INTRODUCTION TO CHATBOTS
Chatbots is a computer program designed to simulate conversation with human users,
especially over the internet powered by Artificial intelligence.
Mitsuku Bot
CleverBot*
Rose*
Types of Chatbots
Simple Chatbot (Script bots)
Smart Chatbots (AI based Smart bots)
HUMAN LANGUAGE VS COMPUTER LANGUAGE
There are rules in human language.
There are nouns, verbs, adverbs, and adjectives.
A word can be a noun at one time and an adjective some other time.
There are rules to provide structure to a language.
Besides the matter of arrangement, there’s also meaning behind the language we use. Human
communication is complex.

Text processing/Data processing, Bag of words, TFIDF, NLTK


The language of computers is Numerical.
We need to convert our language to numbers.
This conversion takes a few steps to happen.
The first step to it is Text Normalisation.
Since human languages are complex, we need to first of all simplify them in order to make sure
that the understanding becomes possible.
Text Normalisation:
helps in cleaning up the textual data in such a way that it comes down to a level where its
complexity is lower than the actual data.
Text normalisation simplifies the text for further processing.
In Text Normalisation, we undergo several steps to normalize the text to a lower level.
Textual data from multiple documents altogether is known as corpus.
Let us take a look at the steps of Text Normalization:
Stemming
In this step, the remaining words are reduced to their root words.
In other words, stemming is the process in which the affixes of words are removed and the
words are converted to their base form.
Note that in stemming, the stemmed words (words which are we get after removing the affixes)
might not be meaningful.
Here in this example as you can see:
healed, healing and healer
all were reduced to heal but studies was reduced to studi after the affix removal which is not a
meaningful word.
Stemming does not take into account if the stemmed word is meaningful or not.
It just removes the affixes hence it is faster.
Lemmatization
Stemming and lemmatization both are alternative processes to each other as the role of both
the processes is same –removal of affixes.
But the difference between both of them is that in lemmatization, the word we get after affix
removal (also known as lemma) is a meaningful one.
Lemmatization makes sure that lemma is a word with meaning and hence it takes a longer time
to execute than stemming.

Bag of words (BOW)


Bag of Words is a Natural Language Processing model which helps in extracting features out of
the text which can be helpful in machine learning algorithms.
In a bag of words, we get the occurrences of each word and construct the vocabulary for the
corpus.

Explain the key steps of NLP – based text analysis.

i) Sentence Segmentation
ii) Tokenization
iii) Removing Stop words, Special Characters and Numbers
iv) Stemming
v) Converting Text to common Case
vi) Lemmatization

What are some of the applications of chatbots in health care ?


The most valuable features of using chatbots in healthcare include:
· Monitoring: Awareness and tracking of user’s behavior, anxiety, and weight changes
to encourage developing better habits.
· Anonymity: Especially in sensitive and mental health issues.
· Personalization: Level of personalization depends on the specific application. Some
applications make use of measurements of:
. Physical vitals (oxygenation, heart rhythm, body temperature) via mobile sensors.
. Patient behavior via facial recognition.
· Real time interaction: Immediate response, notifications, and reminders.
· Scalability: Ability to react with numerous users at the same time.

4. Explain the difference between Stemming and Lemmatization.

Stemming: Stemming is a rudimentary rule-based process of stripping the suffixes


(“ing”, “ly”, “es”, “s” etc) from a word.

154

Stemming is a process of reducing words to their word stem, base or root form (for
example, books — book, looked — look).

Lemmatization: Lemmatization, on the other hand, is an organized & step by step


procedure of obtaining the root form of the word, it makes use of vocabulary
(dictionary importance of words) and morphological analysis (word structure and
grammar relations).
5. What is the difference between how humans interpret communication and how NLP interpret?
The communications made by the machines are very basic and simple. Human communication
is complex. There are multiple characteristics of the human language that might be easy for a
human to understand but extremely difficult for a computer to understand.
For machines it is difficult to understand our language. Let us take a look at some of them
here:
Arrangement of the words and meaning - There are rules in human language. There are nouns,
verbs, adverbs, adjectives. A word can be a noun at one time and an adjective some other time.
This can create difficulty while processing by computers.
Analogy with programming language- Different syntax, same semantics: 2+3 = 3+2 Here the
way these statements are written is different, but their meanings are the same that is 5.
Different semantics, same syntax: 2/3 (Python 2.7) ≠ 2/3 (Python 3) Here the statements
written have the same syntax but their meanings are different. In Python 2.7, this statement
would result in 1 while in Python 3, it would give an output of 1.5. Multiple Meanings of a
word - In natural language, it is important to understand that a word can have multiple
meanings and the meanings fit into the statement according to the context of it.

Bag of Words is a Natural Language Processing model which helps in extracting


features out of the text which can be helpful in machine learning algorithms. In bag of
words, we get the occurrences of each word and construct the vocabulary for the
corpus. Bag of Words just creates a set of vectors containing the count of word
occurrences in the document (reviews). Bag of Words vectors are easy to interpret.

https://drive.google.com/file/d/1umULIE7hDLTYA9JDOyJrqIUmSeWIrX51/view

You might also like