Paper Review
Paper Review
Presented By:
Aayushee Gupta
Supervised By:
Prof. G.Srinivasaraghavan
• Discussed current state of art and future of AI
• Whether or not an AI system truly understands what it
does
• What all AI lacks: adaptability, robustness, abstraction,
AI Debate 2 generalizability, common sense, and causal reasoning
• Importance of building hybrid systems
• Neuro v/s Symbolic AI Neuro-Symbolic AI
Thinking Fast
and Slow
• Human decisions are guided
by two main capabilities:
System 1 : Fast, intuitive and
unconscious thinking
System 2 : Slow, rational and
logical thinking
• Loose comparison between
System 1 and Machine Learning
System 2 and Symbolic Logic Learning
Deep
Learning Two most fundamental aspects of intelligent cognitive behaviour:
the ability to learn from experience and the ability to reason from
Criticism what has been learned – Leslie Valiant
Develop a framework for building systems that can routinely
acquire, represent, and manipulate abstract knowledge, with a
focus on building systems that use that knowledge in the service
of building, updating, and reasoning over complex, internal models
of the external world – Gary Marcus
Advantages of Neuro-Symbolic Integration
Reasoning takes place either symbolically or within the network in distributed form
Neural network-based learning and inference under uncertainty have been expected
to address the brittleness and computational complexity of symbolic systems
• Key Questions:
Attention layer or graph networks?
Whether to use probability theory and at what level
Does theorem proving using neural networks bring a significant gain?
Localist v/s distributed representation
Deep networks or Bayesian networks?
• Ingredients:
Gradient-based optimization used by deep learning to handle large amounts of data
Language for describing encoded knowledge: First order logic, nonmonotonic and modal logic and logic
programming
Modularity: refer to large parts of the network by the composition of symbols and relations among them
Reasoning, within or outside the network, exact or approximate
Constraint satisfaction as part of the interplay between learning and reasoning
Ways of hybrid integration
learning is carried out from data by neural networks which use gradient descent optimization
efficient forms of propositional reasoning can also be carried out by the network, c.f. neural-
symbolic cognitive reasoning
rich first-order logic reasoning and extrapolation needs to be done symbolically from
descriptions extracted from the trained network
once symbolic meaning has emerged from the trained network, symbols can be manipulated
easily by the current computer and can serve as constraints for further learning from data
Challenges for the third wave in AI
• First order logic and higher order knowledge extraction from very large
networks that is provably sound and efficient
• Goal-directed commonsense and efficient combinatorial reasoning
• Human-network communication for promoting communication and
argumentation
• How symbolic meaning emerges from large networks of neurons
• Proofs are needed of the capability of different neural architectures at
representing various logical languages.
• Setting up standard benchmarks and associated comprehensibility tests
for comparative evaluation in the next decade
Thinking Fast and Slow: Efficient Text-to-Visual
Retrieval with Transformers
• Given a text string, how would you search for related images and videos on a large scale?
• The objective is to learn image embeddings f(x) and text embeddings g(y) so that
semantically related images and text have high similarity and the similarity of unrelated
images and text is low
• Noise Contrastive Estimation Loss:
• The image encoder (f) can be a globally pooled output of a CNN while the text encoder
(g) is either a bag-of-words representation or a more sophisticated BERT encoder
Slow Model : Cross-attention
• Given image 𝓍 and text , following model computes similarity between the
two:
where is the model, is a visual encoder and is the cross-attention
network that computes similarity between text and image
• Fine-grained cross-attention: Gradually upsample the last convolutional
feature map conditioned on earlier higher resolution feature maps
Slow Model : Details
• Serafini, Luciano, and Artur d'Avila Garcez. "Logic tensor networks: Deep learning and logical reasoning
from data and knowledge." arXiv preprint arXiv:1606.04422 (2016).
• Tran, Son N., and Artur S. d’Avila Garcez. "Deep logic networks: Inserting and extracting knowledge from
deep belief networks." IEEE transactions on neural networks and learning systems 29.2 (2016): 246-258.
• Tran, Son N., and Artur S. d’Avila Garcez. "Deep logic networks: Inserting and extracting knowledge from
deep belief networks." IEEE transactions on neural networks and learning systems 29.2 (2016): 246-258.
• Rocktäschel, Tim, and Sebastian Riedel. "End-to-end differentiable proving." Advances in neural
information processing systems 30 (2017).
• Daniel, Kahneman. "Thinking, fast and slow." (2017).
• Manhaeve, Robin, et al. "Deepproblog: Neural probabilistic logic programming." Advances in Neural
Information Processing Systems 31 (2018).
• Booch, Grady, et al. "Thinking fast and slow in AI." arXiv preprint arXiv:2010.06002 (2020).
• Garcez, Artur d'Avila, and Luis C. Lamb. "Neurosymbolic AI: the 3rd wave." arXiv preprint arXiv:2012.05876 (2020).
• Miech, Antoine, et al. "Thinking fast and slow: Efficient text-to-visual retrieval with transformers." Proceedings of
the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021.