AI Questions and Answers
AI Questions and Answers
1) What is AI?
- AI is a form of intelligence; a type of technology and a field of study. AI theory and development of computer
systems (both machines and software) are able to perform tasks that normally require human intelligence.
Artificial Intelligence covers a broad range of domains and applications and is expected to impact every field in
the future. Overall, its core idea is building machines and algorithms which are capable of performing
computational tasks that would otherwise require human like brain functions.
- Resources to refer:
- https://www.weforum.org/platforms/shaping-the-future-of-technology-governance-artificial-
intelligence-and-machine-learning
- https://www.britannica.com/technology/artificial-intelligence
- http://www.europarl.europa.eu/RegData/etudes/STUD/2018/626074/IPOL_STU(2018)626074_EN.pdf
- https://niti.gov.in/writereaddata/files/document_publication/NationalStrategy-for-AI-Discussion-
Paper.pdf
- Depending upon datasets, various applications of Artificial Intelligence exists in today’s time. Some of them are:
- Data Science:
- Prediction of Air Quality
- Weather Forecasting
- Netflix
- Amazon Shopping, etc.
- Computer Vision
- Snapchat
- Image Editing Apps
- Face Lock
- Identity Recognition, etc.
- Natural Language Processing
- Auto-correct
- Text-to-Speech
- Speech Recognition
- Twitter, etc.
4) Define: AI,ML and DL
- Artificial Intelligence, or AI, refers to any technique that enables computers to mimic human intelligence.
The AI-enabled machines think algorithmically and execute what they have been asked for intelligently.
- Machine Learning, or ML, enables machines to improve at tasks with experience. The machine learns from
its mistakes and takes them into consideration in the next execution. It improvises itself using its own
experiences.
- Deep Learning, or DL, enables software to train itself to perform tasks with vast amounts of data. In deep
learning, the machine is trained with huge amounts of data which helps it into training itself around the
data. Such machines are intelligent enough to develop algorithms for themselves.
5) What are the different types of Approaches taken for building AI?
- Neural networks are loosely modelled after how neurons in the human brain behave.
- Neural networks come under Deep Learning as they are used for efficient results for vast datasets.
- The key advantage of neural networks, is that they are able to extract data features automatically without needing
the input of the programmer.
- A neural network is essentially a system of organizing machine learning algorithms to perform certain tasks.
- It is a fast and efficient way to solve problems for which the dataset is very large, such as in images.
8) What is Python?
9) Why Python?
- Python is an open-source language which makes it easier for everyone to access and work upon.
- Since it is free to use, a lot of people work around Python hence there exists many packages and modules which
can directly be used while programming in Python. One does not need to write everything from scratch.
- The language is simple. There isn’t much syntax which needs to be learnt and it is also very user-friendly.
- A lot of tutorials and courses are available online for free to learn python. Many of them offer basic working with
Python and a few also talk about advance learnings around it.
- Some of the resources are:
- https://docs.python.org/3/tutorial/
- https://www.tutorialspoint.com/python/index.htm
- https://www.datacamp.com/courses/intro-to-python-for-data-science
- https://www.w3schools.com/python/
- The Jupyter Notebook is an open source application that you can use to create and share documents that contain
live code, equations, visualizations, and text.
- A lot of interfaces are available when it comes to working on Python. Jupyter Notebook is prefered because of its
user-friendliness.
- In a jupyter notebook, one can work around the combination of all - Text, Images, Live Code, etc.
- The understandability of concepts in Jupyter notebook becomes better.
13) How to use Jupyter Notebook?
DATA SCIENCE
1) What is AI for Data Science?
- Data science is the study of data. It involves developing methods of recording, storing, and analyzing data to
effectively extract useful information.
- The goal of data science is to gain insights and knowledge from any type of data — both structured and
unstructured.
- AI for data Science is the study of data, collection and analysis, etc. in order to predict important information out
of it.
- Data Science is a very versatile field as its applications are quite diverse. Few applications are mentioned below:
- Targeted Advertising: Showing exact advertisements to user about what he wants.
- Google Maps: Navigating through the shortest route possible
- Weather Forecasting: Predicting the data for a short period of time by analysing weather patterns
- Recommender engines: Recommending what the user wants depending upon their choices
- Predictions: Predicting the result of a process on the basis of several previous outcomes
3) What are the basic statistical terms used for Data Analysis?
- Basic statistics are used for Data Analysis. Some of the functions used are:
- Mean: Mean is the sum of all the data points divided by the number of data points.
- Median: Median is the 50th percentile value and is useful if the distribution is skewed.
- Mode: Mode is the most frequent value and it can help us identify the most frequent occurence.
- Standard Deviation: Standard deviation is a measure of how spread out the numbers are and gives a
standard way of knowing what is normal. It is the square root of variance.
- Variance: Variance is the average of the squared differences from the mean.
- Various graphical representations can be used for data analysis. Some of them are mentioned below:
- Box plot
- Histogram
- Scatter plot
- Line graph
- Bar graph
5) What are packages? Which packages are commonly used in Data Science?
- A package is nothing but a space where you can find codes or functions or modules of similar type. There are
various packages readily available to use for free (perks of python being an open-source language) for various
purposes.
- Some of the packages commonly used for Data Sciences are:
- NumPy
- Matplotlib
- Pandas
6) What is NumPy?
- NumPy, which stands for Numerical Python, is the fundamental package for scientific computing in Python
consisting of high-performance multidimensional array objects and a collection of tools for working with those
arrays. Using NumPy, mathematical and logical operations on arrays can be performed.
- Read more about NumPy here:
- https://www.tutorialspoint.com/numpy/index.htm
- https://www.geeksforgeeks.org/numpy-in-python-set-1-introduction/
- https://www.datacamp.com/community/tutorials/python-numpy-tutorial
7) What is Matplotlib.PyPlot?
- Matplotlib is a plotting library for the Python programming language and its numerical mathematics extension
NumPy. It is a huge package which contains a lot of extensions to support other tools and packages in Python.
Here, for basic data science implementations, we use one of its attributes called PyPlot.
- Matplotlib.PyPlot is a plotting library used for 2D graphics in python programming language.
- We use Matplotlib.PyPlot to visualize data in graphical form to understand patterns, trends, abnormalities,
outliers,etc. Which also helps us understand which models to use for prediction.
- Read more about Matplotlib.PyPlot here:
- https://www.geeksforgeeks.org/python-introduction-matplotlib/
- https://matplotlib.org/3.1.1/tutorials/introductory/pyplot.html
- https://www.tutorialspoint.com/numpy/numpy_matplotlib.htm
8) What is Pandas?
- Pandas is an open source package providing high-performance, easy-to-use data structures and data analysis tools
for the python programming language.
- It is the most popular python library that is used for data analysis. It provides highly optimized performance with
back-end source code is purely written in Python.
- Read more about Pandas here:
- https://www.geeksforgeeks.org/python-data-analysis-using-pandas/
- https://datacamp.com/community/tutorials/pandas-tutorial-dataframe-python
- https://www.tutorialspoint.com/python_pandas/index.htm
COMPUTER VISION
- Computer vision is a field of computer science that works on enabling computers to see, identify and process
images in the same way that human vision does, and then provide appropriate output.
- Read more about this here:
- https://en.wikipedia.org/wiki/Computer_vision
- https://www.techopedia.com/definition/32309/computer-vision
3) Mention some of the image processing techniques used for Computer Vision.
4) What is a pixel?
- A pixel is the smallest unit of a digital image or graphic that can be displayed and represented on a digital display
device. A pixel is the basic logical unit in digital graphics. Pixels are combined to form a complete image, video,
text or any visible thing on a computer display.
- Read more about pixels here: https://www.techopedia.com/definition/24012/pixel
- A pixel value can vary from 0 to 255, where 0 represents no colour or black and 255 represents full colour or
white.
- Grayscale image only has shades of black and white. Gray scale image is usually 8-bit image with each pixel having
256 combinations of shades of gray. It can be represented using just one plane/channel of pixels.
- Whereas color image usually is a 24 bit image with 8-bits of Red, 8-bits of Green, 8-bits for blue information. It
can be represented using three planes/ channels one each of Red, Green and Blue.
- For more reading, refer to the resources below:
- http://aishack.in/tutorials/color-spaces-1/
- https://www.allaboutcircuits.com/technical-articles/understanding-color-models-used-in-digital-image-
processing/
7) What is OpenCV?
- OpenCV (Open Source Computer Vision Library) is an open source computer vision and machine learning
software library which provides the user with various image-processing tools which can be easily used for any
Computer Vision application.
- To read more, refer to this link: https://opencv.org/about/
8) What are the basic functions in OpenCV available for Image Processing?
- Natural Language Processing (or NLP) is a domain of Artificial Intelligence which works around speech and textual
data. The AI models working in the domain of NLP focus majorly on text interpretation, speech recognition,
meaning extraction, etc.
- A chatbot is an artificial intelligence (AI) software that can simulate a conversation (or chat) with a user in natural
language through messaging applications, websites, mobile apps or through the telephone.
- A chatbot is often described as one of the most advanced and promising expressions of interaction between
humans and machines. However, from a technological point of view, a chatbot only represents the natural
evolution of a Question Answering system leveraging Natural Language Processing (NLP). Formulating responses
to questions in natural language is one of the most typical Examples of Natural Language Processing applied in
various enterprises’ end-user applications.
4) Mention some examples of chatbots.
5) What are the tasks which could be performed under Natural Language Processing?
- Since Natural Language Processing majorly deals with text processing, following are some operations that are
performed under NLP:
- Sentence Segmentation: Dividing the whole text into sentences.
- Tokenization: Separating sentences into words, numbers and special characters.
- Removing Stop words: Cleaning textual data by removing words with minimal importance.
- Case Conversion: Converting the whole data in same case (preferably lower case)
- Stemming: Reducing words to their root form (example: stepping, stepped, stepper becomes step)
irrespective of root word being a meaningful word.
- Lemmatization: Reducing words to their root form (same as stemming) making sure that the root word is
a meaningful one.
- Bag of words: Creating vocabulary for the textual data and calculating the occurence of words in the
dataset.
- TFIDF: Term Frequency - Inverse Document Frequency is the tool through which we can understand the
value of a word in the whole dataset.
- Read more about these tools here:
- https://www.geeksforgeeks.org/text-preprocessing-in-python-set-1/
- https://medium.com/@datamonsters/text-preprocessing-in-python-steps-tools-and-examples-
bf025f872908
- Natural Language Tool-Kit or NLTK is a package readily available for text-processing in Python.
- The package contains functions and modules which can be used for Natural Language Processing.
- Read further about the package here:
- https://www.nltk.org/
- https://pythonspot.com/category/nltk/
- https://www.datacamp.com/community/tutorials/text-analytics-beginners-nltk