0% found this document useful (0 votes)
66 views17 pages

Ver 3

informative

Uploaded by

Durga Ganesan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
66 views17 pages

Ver 3

informative

Uploaded by

Durga Ganesan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 17

VERTICAL 7: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

CCS350 KNOWLEDGE ENGINEERING LTP


C202
3
COURSE OBJECTIVES:
• To understand the basics of Knowledge Engineering.
• To discuss methodologies and modeling for Agent Design and Development.
• To design and develop ontologies.
• To apply reasoning with ontologies and rules.
• To understand learning and rule learning.

UNIT I REASONING UNDER UNCERTAINTY 6


Introduction – Abductive reasoning – Probabilistic reasoning: Enumerative Probabilities – Subjective
Bayesian view – Belief Functions – Baconian Probability – Fuzzy Probability – Uncertainty methods
- Evidence-based reasoning – Intelligent Agent – Mixed-Initiative Reasoning – Knowledge
Engineering.

UNIT II METHODOLOGY AND MODELING 6


Conventional Design and Development – Development tools and Reusable Ontologies – Agent
Design and Development using Learning Technology – Problem Solving through Analysis and
Synthesis – Inquiry-driven Analysis and Synthesis – Evidence-based Assessment – Believability
Assessment – Drill-Down Analysis, Assumption-based Reasoning, and What-If Scenarios.

UNIT III ONTOLOGIES – DESIGN AND DEVELOPMENT 6


Concepts and Instances – Generalization Hierarchies – Object Features – Defining Features –
Representation – Transitivity – Inheritance – Concepts as Feature Values – Ontology Matching.
Design and Development Methodologies – Steps in Ontology Development – Domain
Understanding and Concept Elicitation – Modelling-based Ontology Specification.

UNIT IV REASONIING WITH ONTOLOGIES AND RULES 6


Production System Architecture – Complex Ontology-based Concepts – Reduction and Synthesis
rules and the Inference Engine – Evidence-based hypothesis analysis – Rule and OntologyMatching
– Partially Learned Knowledge – Reasoning with Partially Learned Knowledge.

UNIT V LEARNING AND RULE LEARNING 6


Machine Learning – Concepts – Generalization and Specialization Rules – Types – Formal
definition of Generalization. Modelling, Learning and Problem Solving – Rule learning and
Refinement – Overview – Rule Generation and Analysis – Hypothesis Learning.
30 PERIODS
PRACTICAL EXERCISES: 30 PERIODS
1. Perform operations with Evidence Based Reasoning.
2. Perform Evidence based Analysis.
3. Perform operations on Probability Based Reasoning.
4. Perform Believability Analysis.
5. Implement Rule Learning and refinement.
6. Perform analysis based on learned patterns.
7. Construction of Ontology for a given domain.

1
COURSE OUTCOMES:
At the end of this course, the students will be able to:
CO1: Understand the basics of Knowledge Engineering.
CO2: Apply methodologies and modelling for Agent Design and Development.
CO3: Design and develop ontologies.
CO4: Apply reasoning with ontologies and rules.
CO5: Understand learning and rule learning.
TOTAL: 60 PERIODS
TEXT BOOKS:
1. Gheorghe Tecuci, Dorin Marcu, Mihai Boicu, David A. Schum, Knowledge Engineering
Building Cognitive Assistants for Evidence-based Reasoning, Cambridge University Press,
First Edition, 2016. (Unit 1 – Chapter 1 / Unit 2 – Chapter 3,4 / Unit 3 – Chapter 5, 6 / Unit
4
- 7 , Unit 5 – Chapter 8, 9 )

2
REFERENCES:
1. Ronald J. Brachman, Hector J. Levesque: Knowledge Representation and Reasoning,
Morgan Kaufmann, 2004.
2. Ela Kumar, Knowledge Engineering, I K International Publisher House, 2018.
3. John F. Sowa: Knowledge Representation: Logical, Philosophical, and Computational
Foundations, Brooks/Cole, Thomson Learning, 2000.
4. King , Knowledge Management and Organizational Learning , Springer, 2009.
5. Jay Liebowitz, Knowledge Management Learning from Knowledge Engineering, 1st
Edition,2001.

CO’s-PO’s & PSO’s MAPPING


PO’s PSO’s
CO’s
1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
1 3 1 1 1 1 1 - - 1 2 1 2 1 1 1
2 3 2 3 2 2 - - - 2 1 2 1 3 3 1
3 2 2 3 2 2 - - - 3 2 2 2 3 2 3
4 2 2 3 1 1 - - - 2 2 2 2 2 1 1
5 2 2 2 1 1 - - - 2 1 1 1 2 1 1
AVg. 2.4 1.8 2.4 1.4 1.4 0.2 0 0 2 1.6 1.6 1.6 2.2 1.6 1.4
1 - low, 2 - medium, 3 - high, ‘-' - no correlation

CCS364 SOFT COMPUTING LTP


C2 0
2 3
COURSE OBJECTIVES:
 To introduce the ideas of fuzzy sets, fuzzy logic and use of heuristics based on
human experience.
 To provide the mathematical background for carrying out the optimization associated
with neural network learning
 To learn various evolutionary Algorithms.
 To become familiar with neural networks that can learn from available examples
and generalize to form appropriate rules for inference systems.
 To introduce case studies utilizing the above and illustrate the Intelligent behavior
of programs based on soft computing

UNIT 1 INTRODUCTION TO SOFT COMPUTING AND FUZZY LOGIC 6


Introduction - Fuzzy Logic - Fuzzy Sets, Fuzzy Membership Functions, Operations on Fuzzy Sets,
Fuzzy Relations, Operations on Fuzzy Relations, Fuzzy Rules and Fuzzy Reasoning,
FuzzyInference Systems

UNIT II NEURAL NETWORKS 6


Supervised Learning Neural Networks – Perceptrons - Backpropagation -Multilayer Perceptrons –
Unsupervised Learning Neural Networks – Kohonen Self-Organizing Networks

3
UNIT III GENETIC ALGORITHMS 6
Chromosome Encoding Schemes -Population initialization and selection methods - Evaluation
function - Genetic operators- Cross over – Mutation - Fitness Function – Maximizing function

UNIT IV NEURO FUZZY MODELING 6


ANFIS architecture – hybrid learning – ANFIS as universal approximator – Coactive Neuro fuzzy
modeling – Framework – Neuron functions for adaptive networks – Neuro fuzzy spectrum -
Analysis of Adaptive Learning Capability

UNIT V APPLICATIONS 6
Modeling a two input sine function - Printed Character Recognition – Fuzzy filtered neural
networks – Plasma Spectrum Analysis – Hand written neural recognition - Soft Computing for
Color Recipe Prediction.
30 PERIODS
COURSE OUTCOMES:
CO1: Understand the fundamentals of fuzzy logic operators and inference mechanisms
CO2: Understand neural network architecture for AI applications such as classification and
clustering
CO3: Learn the functionality of Genetic Algorithms in Optimization
problems
CO4: Use hybrid techniques involving Neural networks and Fuzzy logic
CO5: Apply soft computing techniques in real world applications

PRACTICAL EXERCISES 30 PERIODS


1. Implementation of fuzzy control/ inference system
2. Programming exercise on classification with a discrete perceptron
3. Implementation of XOR with backpropagation algorithm
4. Implementation of self organizing maps for a specific application
5. Programming exercises on maximizing a function using Genetic algorithm
6. Implementation of two input sine function
7. Implementation of three input non linear function
TOTAL: 60 PERIODS
TEXT BOOKS:
1. SaJANG, J.-S. R., SUN, C.-T., & MIZUTANI, E. (1997). Neuro-fuzzy and soft
computing: Acomputational approach to learning and machine intelligence. Upper Saddle
River, NJ, Prentice Hall,1997
2. Himanshu Singh, Yunis Ahmad Lone, Deep Neuro-Fuzzy Systems with Python
3. With Case Studies and Applications from the Industry, Apress, 2020

REFERENCES
1. roj Kaushik and Sunita Tiwari, Soft Computing-Fundamentals Techniques
and Applications,1st Edition, McGraw Hill, 2018.
2. S. Rajasekaran and G.A.V.Pai, “Neural Networks, Fuzzy Logic and Genetic
Algorithms”, PHI,2003.

4
3. Samir Roy, Udit Chakraborthy, Introduction to Soft Computing, Neuro Fuzzy and
GeneticAlgorithms, Pearson Education, 2013.
4. S.N. Sivanandam, S.N. Deepa, Principles of Soft Computing, Third Edition, Wiley India
PvtLtd, 2019.
5. R.Eberhart, P.Simpson and R.Dobbins, “Computational Intelligence - PC Tools”, AP
Professional, Boston, 1996

CO’s-PO’s & PSO’s MAPPING


PO’s PSO’s
CO’s
1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
1 3 2 3 3 3 - - - 3 1 3 2 3 1 2
2 2 3 3 2 3 - - - 3 2 3 2 2 1 3
3 1 3 2 2 1 - - - 3 1 1 2 1 3 2
4 1 2 1 3 2 - - - 3 3 1 1 2 1 1
5 2 3 1 2 1 - - - 3 3 3 2 1 2 3
AVg. 1.8 2.6 2 2.4 2 - - - 3 2 2.2 1.8 1.8 1.6 2.2
1 - low, 2 - medium, 3 - high, ‘-' - no correlation

CCS355 NEURAL NETWORKS AND DEEP LEARNING L T PC


2 0 2 3
COURSE OBJECTIVES:
 To understand the basics in deep neural networks
 To understand the basics of associative memory and unsupervised learning networks
 To apply CNN architectures of deep neural networks
 To analyze the key computations underlying deep learning, then use them to build and
train deep neural networks for various tasks.
 To apply autoencoders and generative models for suitable applications.

UNIT I INTRODUCTION 6
Neural Networks-Application Scope of Neural Networks-Artificial Neural Network: An
Introduction- Evolution of Neural Networks-Basic Models of Artificial Neural Network- Important
Terminologies of ANNs-Supervised Learning Network.

UNIT II ASSOCIATIVE MEMORY AND UNSUPERVISED LEARNING NETWORKS 6


Training Algorithms for Pattern Association-Autoassociative Memory Network-Heteroassociative
Memory Network-Bidirectional Associative Memory (BAM)-Hopfield Networks-Iterative
Autoassociative Memory Networks-Temporal Associative Memory Network-Fixed Weight
Competitive Nets-Kohonen Self-Organizing Feature Maps-Learning Vector Quantization-Counter
propagation Networks-Adaptive Resonance Theory Network.

UNIT III THIRD-GENERATION NEURAL NETWORKS 6


Spiking Neural Networks-Convolutional Neural Networks-Deep Learning Neural Networks-
Extreme Learning Machine Model-Convolutional Neural Networks: The Convolution Operation –
Motivation –Pooling – Variants of the basic Convolution Function – Structured Outputs – Data
Types – Efficient Convolution Algorithms – Neuroscientific Basis – Applications: Computer Vision,
Image Generation,

5
Image Compression.

UNIT IV DEEP FEEDFORWARD NETWORKS 6


History of Deep Learning- A Probabilistic Theory of Deep Learning- Gradient Learning – Chain
Rule and Backpropagation - Regularization: Dataset Augmentation – Noise Robustness -Early
Stopping, Bagging and Dropout - batch normalization- VC Dimension and Neural Nets.

UNIT V RECURRENT NEURAL NETWORKS 6


Recurrent Neural Networks: Introduction – Recursive Neural Networks – Bidirectional RNNs –
Deep Recurrent Networks – Applications: Image Generation, Image Compression, Natural
Language Processing. Complete Auto encoder, Regularized Autoencoder, Stochastic Encoders and
Decoders, Contractive Encoders.

30 PERIODS
LAB EXPERIMENTS: 30 PERIODS
1. Implement simple vector addition in TensorFlow.
2. Implement a regression model in Keras.
3. Implement a perceptron in TensorFlow/Keras Environment.
4. Implement a Feed-Forward Network in TensorFlow/Keras.
5. Implement an Image Classifier using CNN in TensorFlow/Keras.
6. Improve the Deep learning model by fine tuning hyper parameters.
7. Implement a Transfer Learning concept in Image Classification.
8. Using a pre trained model on Keras for Transfer Learning
9. Perform Sentiment Analysis using RNN
10. Implement an LSTM based Autoencoder in TensorFlow/Keras.
11. Image generation using GAN
Additional Experiments:
1. Train a Deep learning model to classify a given image using pre trained model
2. Recommendation system from sales data using Deep Learning
3. Implement Object Detection using CNN
4. Implement any simple Reinforcement Algorithm for an NLP problem
TOTAL: 60 PERIODS
COURSE OUTCOMES:
At the end of this course, the students will be able to: CO1:
Apply Convolution Neural Network for image processing.
CO2: Understand the basics of associative memory and unsupervised learning networks.
CO3: Apply CNN and its variants for suitable applications.
CO4: Analyze the key computations underlying deep learning and use them to build and train deepneural
networks for various tasks.
CO5: Apply autoencoders and generative models for suitable applications.

6
TEXT BOOKS:
1. Ian Goodfellow, Yoshua Bengio, Aaron Courville, “Deep Learning”, MIT Press, 2016.
2. Francois Chollet, “Deep Learning with Python”, Second Edition,
Manning Publications,2021.

REFERENCES:
1. Aurélien Géron, “Hands-On Machine Learning with Scikit-Learn and TensorFlow”,
Oreilly,2018.
2. Josh Patterson, Adam Gibson, “Deep Learning: A Practitioner’s Approach”,
O’Reilly Media,2017.
3. Charu C. Aggarwal, “Neural Networks and Deep Learning: A Textbook”, Springer
International Publishing, 1st Edition, 2018.
4. Learn Keras for Deep Neural Networks, Jojo Moolayil, Apress,2018
5. Deep Learning Projects Using TensorFlow 2, Vinita Silaparasetty, Apress, 2020
6. Deep Learning with Python, FRANÇOIS CHOLLET, MANNING SHELTER ISLAND,2017.
7. S Rajasekaran, G A Vijayalakshmi Pai, “Neural Networks, FuzzyLogic and Genetic
Algorithm, Synthesis and Applications”, PHI Learning, 2017.
8. Pro Deep Learning with TensorFlow, Santanu Pattanayak, Apress,2017
9. James A Freeman, David M S Kapura, “Neural Networks Algorithms, Applications, and
Programming Techniques”, Addison Wesley, 2003.
CO’s-PO’s & PSO’s MAPPING
PO’s PSO’s
CO’s
1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
1 3 2 3 2 3 1 - - 2 1 - - 2 2 1
2 3 1 2 1 - - - - - 1 2 2 - 1 -
3 3 3 3 3 3 1 - - 2 1 - - 2 2 1
4 3 3 3 3 3 - - - 2 - 2 3 2 2 2
5 1 1 3 2 3 - - - 2 - - - 1 1 -
AVg. 2.6 2 2.8 2.2 2.4 0.4 0 0 1.6 0.6 0.8 1 1.4 1.6 0.8
1 - low, 2 - medium, 3 - high, ‘-' - no correlation

CCS369 TEXT AND SPEECH ANALYSIS LTP


C2 0
2 3
COURSE OBJECTIVES:
 Understand natural language processing basics
 Apply classification algorithms to text documents
 Build question-answering and dialogue systems
 Develop a speech recognition system
 Develop a speech synthesizer

UNIT I NATURAL LANGUAGE BASICS 6


Foundations of natural language processing – Language Syntax and Structure- Text
Preprocessingand Wrangling – Text tokenization – Stemming – Lemmatization – Removing stop-
words – Feature Engineering for Text representation – Bag of Words model- Bag of N-Grams
model – TF-IDF model

7
Suggested Activities
● Flipped classroom on NLP
● Implementation of Text Preprocessing using NLTK
● Implementation of TF-IDF models

Suggested Evaluation Methods


Quiz on NLP Basics
Demonstration of Programs

UNIT II TEXT CLASSIFICATION 6


Vector Semantics and Embeddings -Word Embeddings - Word2Vec model – Glove model –
FastText model – Overview of Deep Learning models – RNN – Transformers – Overview of Text
summarization and Topic Models
Suggested Activities
 Flipped classroom on Feature extraction of documents
 Implementation of SVM models for text classification
 External learning: Text summarization and Topic models

Suggested Evaluation Methods


 Assignment on above topics
 Quiz on RNN, Transformers
 Implementing NLP with RNN and Transformers

UNIT III QUESTION ANSWERING AND DIALOGUE SYSTEMS 9


Information retrieval – IR-based question answering – knowledge-based question answering –
language models for QA – classic QA models – chatbots – Design of dialogue systems -–
evaluating dialogue systems

Suggested Activities:
 Flipped classroom on language models for QA
 Developing a knowledge-based question-answering system
 Classic QA model development

Suggested Evaluation Methods


 Assignment on the above topics
 Quiz on knowledge-based question answering system
 Development of simple chatbots

UNIT IV TEXT-TO-SPEECH SYNTHESIS 6


Overview. Text normalization. Letter-to-sound. Prosody, Evaluation. Signal processing -
Concatenative and parametric approaches, WaveNet and other deep learning-based TTSsystems

Suggested Activities:
 Flipped classroom on Speech signal processing
 Exploring Text normalization
 Data collection
 Implementation of TTS systems

8
Suggested Evaluation Methods
 Assignment on the above topics
 Quiz on wavenet, deep learning-based TTS systems
 Finding accuracy with different TTS systems

UNIT V AUTOMATIC SPEECH RECOGNITION 6


Speech recognition: Acoustic modelling – Feature Extraction - HMM, HMM-DNN systems

Suggested Activities:
 Flipped classroom on Speech recognition.
 Exploring Feature extraction

Suggested Evaluation Methods


 Assignment on the above topics
 Quiz on acoustic modelling
30 PERIODS
PRACTICAL EXERCISES 30 PERIODS
10. Create Regular expressions in Python for detecting word patterns and tokenizing text
11. Getting started with Python and NLTK - Searching Text, Counting
Vocabulary, FrequencyDistribution, Collocations, Bigrams
12. Accessing Text Corpora using NLTK in Python
13. Write a function that finds the 50 most frequently occurring words of a text that are not
stop words.
14. Implement the Word2Vec model
15. Use a transformer for implementing classification
16. Design a chatbot with a simple dialog system
17. Convert text to speech and find accuracy
18. Design a speech recognition system and find the error rate
TOTAL: 60 PERIODS
COURSE OUTCOMES:
On completion of the course, the students will be able to
CO1:Explain existing and emerging deep learning architectures for text and speech processing
CO2:Apply deep learning techniques for NLP tasks, language modelling and machine translation
CO3:Explain coreference and coherence for text processing
CO4:Build question-answering systems, chatbots and dialogue systems
CO5:Apply deep learning models for building speech recognition and text-to-speech systems

TEXTBOOK
1. Daniel Jurafsky and James H. Martin, “Speech and Language Processing: An Introduction to
Natural Language Processing, Computational Linguistics, and Speech Recognition”, Third
Edition, 2022.
REFERENCES:
1. Dipanjan Sarkar, “Text Analytics with Python: A Practical Real-World approach to
GainingActionable insights from your data”, APress,2018.
2. Tanveer Siddiqui, Tiwary U S, “Natural Language Processing and Information Retrieval”,
Oxford University Press, 2008.

9
3. Lawrence Rabiner, Biing-Hwang Juang, B. Yegnanarayana, “Fundamentals of Speech
Recognition” 1st Edition, Pearson, 2009.
4. Steven Bird, Ewan Klein, and Edward Loper, “Natural language processing with Python”,
O’REILLY.

CO’s-PO’s & PSO’s MAPPING


PO’s PSO’s
CO’s
1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
1 3 2 3 1 3 - - - 1 2 1 2 1 1 1
2 3 1 2 1 3 - - - 2 2 1 3 3 2 1
3 2 2 1 3 1 - - - 3 3 1 2 3 3 1
4 2 1 1 1 2 - - - 2 1 2 2 3 1 1
5 1 3 2 2 1 - - - 3 2 1 1 2 3 1
AVg. 2.2 1.8 1.8 1.6 2 - - - 2.2 2 1.2 2 2.4 2 1
1 - low, 2 - medium, 3 - high, ‘-' - no correlation

CCS357 OPTIMIZATION TECHNIQUES LT P C


2 0 2
3
COURSE OBJECTIVES:
The objective of this course is to enable the student to
 Formulate and solve linear programming problems (LPP)
 Evaluate Integer Programming Problems, Transportation and Assignment Problems.
 Obtain a solution to network problems using CPM and PERT techniques.
 Able to optimize the function subject to the constraints.
 Identify and solve problems under Markovian queuing models.

UNIT I LINEAR MODELS 6


Introduction of Operations Research - mathematical formulation of LPP- Graphical Methods to
solve LPP- Simplex Method- Two-Phase method

UNIT II INTEGER PROGRAMMING AND TRANSPORTATION PROBLEMS 6


Integer programming: Branch and bound method- Transportation and Assignment problems -
Traveling salesman problem.

UNIT III PROJECT SCHEDULING 6


Project network -Diagram representation – Floats - Critical path method (CPM) – PERT- Cost
considerations in PERT and CPM.

UNIT IV CLASSICAL OPTIMIZATION THEORY 6


Unconstrained problems – necessary and sufficient conditions - Newton-Raphson method,
Constrained problems – equality constraints – inequality constraints - Kuhn-Tucker conditions.

UNIT V QUEUING MODELS 6


Introduction, Queuing Theory, Operating characteristics of a Queuing system, Constituents of a
Queuing system, Service facility, Queue discipline, Single channel models, multiple service
channels.

10
PRACTICALS
1. Solving simplex maximization problems using R programming.
2. Solving simplex minimization problems using R programming.
3. Solving mixed constraints problems – Big M & Two phase method using TORA.
4. Solving transportation problems using R.
5. Solving assignment problems using R.
6. Solving optimization problems using LINGO.
7. Studying Primal-Dual relationships in LP using TORA.
8. Solving LP problems using dual simplex method using TORA.
9. Sensitivity & post optimality analysis using LINGO
10. Solving shortest route problems using optimization software
11. Solving Project Management problems using optimization software
12. Testing random numbers and random variates for their uniformity.
13. Testing random numbers and random variates for their independence
14. Solve single server queuing model using simulation software package.
15. Solve multi server queuing model using simulation software package.
30 PERIODS
COURSE OUTCOMES:
On successful completion of this course, the student will able to
CO1:Formulate and solve linear programming problems (LPP)
CO2:Evaluate Integer Programming Problems, Transportation and Assignment Problems.
CO3:Obtain a solution to network problems using CPM and PERT techniques.
CO4:Able to optimize the function subject to the constraints.
CO5:Identify and solve problems under Markovian queuing models

TEXT BOOK:
1. Hamdy A Taha, Operations Research: An Introduction, Pearson, 10th Edition, 2017.

REFERENCES:
1. ND Vohra, Quantitative Techniques in Management, Tata McGraw Hill, 4th Edition, 2011.
2. J. K. Sharma, Operations Research Theory and Applications, Macmillan, 5th Edition, 2012.
3. Hiller F.S, Liberman G.J, Introduction to Operations Research, 10th Edition McGraw
Hill,2017.
4. Jit. S. Chandran, Mahendran P. Kawatra, KiHoKim, Essentials of Linear
Programming, Vikas Publishing House Pvt.Ltd. New Delhi, 1994.
5. Ravindran A., Philip D.T., and Solberg J.J., Operations Research, John Wiley, 2nd Edition,
2007.
CO’s-PO’s & PSO’s MAPPING
PO’s PSO’s
CO’s 1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
1 3 3 2 1 1 - - - 2 1 1 2 3 3 3
2 3 1 2 2 3 - - - 3 2 3 1 2 1 1
3 2 3 3 2 2 - - - 3 3 1 3 1 3 1
4 2 2 1 1 3 - - - 2 1 3 1 2 1 2
5 2 1 1 3 2 - - - 3 3 1 3 3 2 1
AVg. 2.4 2 1.8 1.8 2.2 - - - 2.6 2 1.8 2 2.2 2 1.6
1 - low, 2 - medium, 3 - high, ‘-' - no correlation

11
CCS348 GAME THEORY LTP
C2 0
2 3
COURSE OBJECTIVES:
 To introduce the student to the notion of a game, its solutions concepts, and other basic
notions and tools of game theory, and the main applications for which they are
appropriate, including electronic trading markets.
 To formalize the notion of strategic thinking and rational choice by using the tools of
game theory, and to provide insights into using game theory in 41odeIIing applications.
 To draw the connections between game theory, computer science, and economics,
especially emphasizing the computational issues.
 To introduce contemporary topics in the intersection of game theory, computer science,
and economics.
 To apply game theory in searching, auctioning and trading.

UNIT I INTRODUCTION 6
Introduction — Making rational choices: basics of Games — strategy — preferences — payoffs —
Mathematical basics — Game theory — Rational Choice — Basic solution concepts-non-
cooperative versus cooperative games — Basic computational issues — finding equilibria and
learning in games- Typical application areas for game theory (e.g. Google's sponsored search, eBay
auctions, electricity trading markets).

UNIT II GAMES WITH PERFECT INFORMATION 6


Games with Perfect Information — Strategic games — prisoner's dilemma, matching pennies -
Nash equilibria —mixed strategy equilibrium — zero-sum games

UNIT III GAMES WITH IMPERFECT INFORMATION 6

Games with Imperfect Information — Bayesian Games — Motivational Examples — General


Definitions — Information aspects — Illustrations — Extensive Games with Imperfect —
Information
— Strategies — Nash Equilibrium —Repeated Games — The Prisoner's Dilemma — Bargaining

UNIT IV NON-COOPERATIVE GAME THEORY 6


Non-cooperative Game Theory — Self-interested agents — Games in normal form — Analyzing
games: from optimality to equilibrium — Computing Solution Concepts of Normal — Form Games
— Computing Nash equilibria of two-player, zero-sum games —Computing Nash equilibria of
two- player, general- sum games — Identifying dominated strategies

UNIT V MECHANISM DESIGN 6


Aggregating Preferences — Social Choice — Formal Model — Voting — Existence of social functions
— Ranking systems — Protocols for Strategic Agents: Mechanism Design — Mechanism design
with unrestricted preferences
30 PERIODS

12
COURSE OUTCOMES:
Upon Completion of the course, the students will be able to
CO1:Discuss the notion of a strategic game and equilibria and identify the characteristics of main
applications of these concepts.
CO2:Discuss the use of Nash Equilibrium for other problems.
CO3:Identify key strategic aspects and based on these be able to connect them to appropriategame
theoretic concepts given a real world situation.
CO4:Identify some applications that need aspects of Bayesian Games.
CO5:Implement a typical Virtual Business scenario using Game theory.

LABORATORY EXERCISES:
1. Prisoner’s dilemma
2. Pure Strategy Nash Equilibrium
3. Extensive Form – Graphs and Trees, Game Trees
4. Strategic Form – Elimination of dominant strategy
5. Minimax theorem, minimax strategies
6. Perfect information games: trees, players assigned to nodes, payoffs, backward Induction,
subgame perfect equilibrium,
7. imperfect-information games - Mixed Strategy Nash Equilibrium - Finding mixed-
strategyNash equilibria for zero sum games, mixed versus behavioral strategies.
8. Repeated Games
9. Bayesian Nash equilibrium

30 PERIODS
TOTAL: 60 PERIODS
TEXT BOOKS:

1. M. J. Osborne, An Introduction to Game Theory. Oxford University Press, 2012.


2. M. Machler, E. Solan, S. Zamir, Game Theory, Cambridge University Press, 2013.
3. N. Nisan, T. Roughgarden, E. Tardos, and V. V. Vazirani, Algorithmic Game
Theory. Cambridge University Press, 2007.
4. A.Dixit and S. Skeath, Games of Strategy, Second Edition. W W Norton & Co Inc, 2004.
5. YoavShoham, Kevin Leyton-Brown, Multiagent Systems: Algorithmic, Game-
Theoretic,and Logical Foundations, Cambridge University Press 2008.
6. Zhu Han, DusitNiyato, WalidSaad, TamerBasar and Are Hjorungnes, “Game Theory in
Wireless and Communication Networks”, Cambridge University Press, 2012.
7. Y.Narahari, “Game Theory and Mechanism Design”, IISC Press, World Scientific.
8. William Spaniel, “Game Theory 101: The Complete Textbook”, CreateSpace Independent
Publishing, 2011.

13
CO’s-PO’s & PSO’s MAPPING
PO’s PSO’s
CO’s
1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
1 3 3 2 3 3 - - - - - - - 1 1 3
2 3 3 3 2 3 - - - - - - - 1 1 1
3 1 1 3 3 3 - - - - - - - 1 1 2
4 2 1 1 1 1 - - - - - - - 1 1 2
5 2 2 3 2 1 - - - - - - - 1 1 2
AVg. 2.2 2 2.4 2.2 2.2 - - - - - - - 1 1 2
1 - low, 2 - medium, 3 - high, ‘-' - no correlation

CCS337 COGNITIVE SCIENCE LTP C


2 023
COURSE OBJECTIVES:
 To know the theoretical background of cognition.
 To understand the link between cognition and computational intelligence.
 To explore probabilistic programming language.
 To study the computational inference models of cognition.
 To study the computational learning models of cognition.

UNIT I PHILOSOPHY, PSYCHOLOGY AND NEUROSCIENCE 6


Philosophy: Mental-physical Relation – From Materialism to Mental Science – Logic and the
Sciences of the Mind – Psychology: Place of Psychology within Cognitive Science – Science of
Information Processing –Cognitive Neuroscience – Perception – Decision – Learning and Memory
– Language Understanding and Processing.

UNIT II COMPUTATIONAL INTELLIGENCE 6


Machines and Cognition – Artificial Intelligence – Architectures of Cognition – Knowledge Based
Systems – Logical Representation and Reasoning – Logical Decision Making –Learning –
Language – Vision.

UNIT III PROBABILISTIC PROGRAMMING LANGUAGE 6


WebPPL Language – Syntax – Using Javascript Libraries – Manipulating probability types and
distributions – Finding Inference – Exploring random computation – Coroutines: Functions that
receive continuations –Enumeration

UNIT IV INFERENCE MODELS OF COGNITION 6


Generative Models – Conditioning – Causal and statistical dependence – Conditional dependence
– Data Analysis – Algorithms for Inference.

UNIT V LEARNING MODELS OF COGNITION 6


Learning as Conditional Inference – Learning with a Language of Thought – Hierarchical
Models–Learning (Deep) Continuous Functions – Mixture Models.
PRACTICAL

14
EXERCISES 30 PERIODS
1. Demonstration of Mathematical functions using WebPPL.
2. Implementation of reasoning algorithms.
3. Developing an Application system using generative model.
4. Developing an Application using conditional inference learning
model.
5. Application development using hierarchical model.
6. Application development using Mixture model.
30 PERIODS
COURSE OUTCOMES:
At the end of this course, the students will be able to:
CO1:Understand the underlying theory behind cognition.
CO2:Connect to the cognition elements computationally.
CO3:Implement mathematical functions through WebPPL.
CO4:Develop applications using cognitive inference Model.
CO5:Develop applications using cognitive learning model.
TOTAL: 60 PERIODS

TEXT BOOK:
1. Vijay V Raghavan,Venkat N.Gudivada, VenuGovindaraju, C.R. Rao, Cognitive
Computing:Theory and Applications: (Handbook of Statistics 35), Elsevier
publications, 2016
2. Judith Hurwitz, Marcia Kaufman, Adrian Bowles, Cognitive Computing and Big Data
Analytics, Wiley Publications, 2015
3. Robert A. Wilson, Frank C. Keil, “The MIT Encyclopedia of the Cognitive
Sciences”,The MITPress, 1999.
4. Jose Luis Bermúdez, Cognitive Science -An Introduction to the Science of the Mind,
Cambridge University Press 2020

REFERENCES:
1. Noah D. Goodman, Andreas Stuhlmuller, “The Design and Implementation of Probabilistic
Programming Languages”, Electronic version of book, https://dippl.org/.
2. Noah D. Goodman, Joshua B. Tenenbaum, The ProbMods Contributors, “Probabilistic
Models of Cognition”, Second Edition, 2016, https://probmods.org/.

CO’s-PO’s & PSO’s MAPPING


PO’s PSO’s
CO’s 1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
1 3 1 3 2 2 - - - 1 1 2 2 1 2 2
2 2 2 1 1 2 - - - 3 2 3 1 2 3 2
3 1 3 1 3 3 - - - 1 3 1 3 3 1 2
4 2 1 1 2 3 - - - 1 2 3 1 3 3 1
5 1 2 3 2 2 - - - 1 2 2 2 2 2 1
AVg 1.8 1.8 1.8 2 2.4 - - - 1.4 2 2.2 1.8 2.2 2.2 1.6
1 - low, 2 - medium, 3 - high, ‘-' - no correlation

15
CCS345 ETHICS AND AI L T P C
2 0 2 3
COURSE OBJECTIVES:
 Study the morality and ethics in AI
 Learn about the Ethical initiatives in the field of artificial intelligence
 Study about AI standards and Regulations
 Study about social and ethical issues of Robot Ethics
 Study about AI and Ethics- challenges and opportunities

UNIT I INTRODUCTION 6
Definition of morality and ethics in AI-Impact on society-Impact on human psychology-Impact on
the legal system-Impact on the environment and the planet-Impact on trust

UNIT II ETHICAL INITIATIVES IN AI 6


International ethical initiatives-Ethical harms and concerns-Case study: healthcare robots,
Autonomous Vehicles , Warfare and weaponization.

UNIT III AI STANDARDS AND REGULATION 6


Model Process for Addressing Ethical Concerns During System Design - Transparency of
Autonomous Systems-Data Privacy Process- Algorithmic Bias Considerations - Ontological
Standard for Ethically Driven Robotics and Automation Systems

UNIT IV ROBOETHICS: SOCIAL AND ETHICAL IMPLICATION OF ROBOTICS 6


Robot-Roboethics- Ethics and Morality- Moral Theories-Ethics in Science and Technology -
Ethical Issues in an ICT Society- Harmonization of Principles- Ethics and Professional
Responsibility- Roboethics Taxonomy.

UNIT V AI AND ETHICS- CHALLENGES AND OPPORTUNITIES 6


Challenges - Opportunities- ethical issues in artificial intelligence- Societal Issues Concerning the
Application of Artificial Intelligence in Medicine- decision-making role in industries-National and
International Strategies on AI.
30 PERIODS

COURSE OUTCOMES:
On completion of the course, the students will be able to
CO1: Learn about morality and ethics in AI
CO2:Acquire the knowledge of real time application ethics, issues and its challenges.
CO3:Understand the ethical harms and ethical initiatives in AI
CO4:Learn about AI standards and Regulations like AI Agent, Safe Design of Autonomous and
Semi-Autonomous Systems
CO5:Understand the concepts of Roboethics and Morality with professional responsibilities.
CO6:Learn about the societal issues in AI with National and International Strategies on AI

PRACTICAL EXERCISES 30 PERIODS


1. Recent case study of ethical initiatives in healthcare, autonomous vehicles and defense
2. Exploratory data analysis on a 2 variable linear regression model

16
3. Experiment the regression model without a bias and with bias
4. Classification of a dataset from UCI repository using a perceptron with and without bias
5. Case study on ontology where ethics is at stake
6. Identification on optimization in AI affecting ethics
TOTAL: 60 PERIODS
TEXT BOOKS :
1. y. Eleanor Bird, Jasmin Fox-Skelly, Nicola Jenner, Ruth Larbey, Emma Weitkamp and Alan
Winfield ,”The ethics of artificial intelligence: Issues and initiatives”, EPRS | European
Parliamentary Research Service Scientific Foresight Unit (STOA) PE 634.452 – March
2020
2. Patrick Lin, Keith Abney, George A Bekey,” Robot Ethics: The Ethical and Social
Implications of Robotics”, The MIT Press- January 2014.
REFERENCES:
1. Towards a Code of Ethics for Artificial Intelligence (Artificial Intelligence:
Foundations,Theory, and Algorithms) by Paula Boddington, November 2017
2. Mark Coeckelbergh,” AI Ethics”, The MIT Press Essential Knowledge series, April 2020
3. Web link:
4. https://sci-hub.mksa.top/10.1007/978-3-540-30301-5_65
5. https://www.scu.edu/ethics/all-about-ethics/artificial-intelligence-and-ethics-sixteen-
challenges-and-opportunities/
6. https://www.weforum.org/agenda/2016/10/top-10-ethical-issues-in-artificial-
intelligence/ 7. https://sci-hub.mksa.top/10.1159/000492428

CO’s-PO’s & PSO’s MAPPING


PO’s PSO’s
CO’s
1 2 3 4 5 6 7 8 9 10 11 12 1 2 3
1 3 2 3 3 1 - - - 1 2 1 1 3 1 1
2 2 1 1 2 1 - - - 1 2 1 1 3 3 1
3 2 3 1 1 3 - - - 2 1 1 2 3 2 2
4 3 1 3 3 2 - - - 2 2 3 1 2 1 3
5 3 1 1 3 3 - - - 2 3 3 3 1 3 3
AVg. 2.6 1.6 1.8 2.4 2 - - - 1.6 2 1.8 1.6 2.4 2 2
1 - low, 2 - medium, 3 - high, ‘-' - no correlation

17

You might also like