Navya Final 1
Navya Final 1
BACHELOR OF TECHNOLOGY
IN
Submitted by
B.Navya Reddy
HT.NO:21D41A6621
Sheriguda, Ibrahimpatam)
(2021-2025)
SRI INDU COLLEGE OF ENGINEERING AND TECHNOLOGY
(An Autonomous Institution under UGC,accredited by NBA, Affiliated to JNTUH
CERTIFICATE
Certified that the Technical Seminar Work entitled “NEURO SYMBOLIC AI FOR NLP” is a
Bonafide work carried out by B.Navya (21D41A6604) in partial fulfillment for the award of
Bachelor of Technology in ARTIFICIAL INTELLIGENCE & MACHINE LEARNING
of SICET, Hyderabad for the academic year 2024-2025.The Technical Seminar Report has been
approved as it satisfies academic requirements in respect of the work prescribed for IV YEAR,
I-SEMISTER of B. TECH course.
1. Introduction 1
3. Core Components 7
7. Explainability in NLP 17
9. Conclusion 21
10. Reference 22
ABSTRACT
Neuro -Symbolic AI is an emerging field that combines the strengths of neural networks and symbolic
reasoning to address complex problems in artificial intelligence. In the context of Natural Language
Processing (NLP), this approach seeks to enhance language understanding by integrating the data-
driven capabilities of neural models with the logical reasoning power of symbolic systems. While
neural networks excel at recognizing patterns and handling large volumes of data, they often lack
transparency and logical consistency. Conversely, symbolic methods are effective for rule-based
reasoning but struggle with scalability and data complexity.
By merging these paradigms, Neuro -Symbolic AI aims to create more interpretable and generalizable
NLP systems. Applications range from natural language understanding, information extraction, and
question answering to explainable AI, where the hybrid approach provides better insights into
decision-making processes. Despite its potential, challenges remain, including the balance between
learning from data and applying logical rules, as well as scaling these systems for real-world
applications. Neuro-Symbolic AI holds significant promise in advancing the field of NLP by
enabling systems that not only process language effectively but also reason about it in a meaningful
way.
INTRODUCTION
Neuro-Symbolic AI is an innovative approach that seeks to combine the best of two major AI paradigms:
neural networks and symbolic reasoning. Neural networks, which form the foundation of modern
machine learning models, are highly effective at learning patterns from vast amounts of data and are
central to the success of deep learning. However, they often act as black boxes, lacking transparency
and the ability to reason explicitly about their decisions. On the other hand, symbolic AI, which uses
logic-based reasoning and rule-based systems, excels in tasks requiring explicit reasoning, such as
logical inference and knowledge representation, but struggles with tasks involving noisy,
unstructured, or large-scale data.
In the field of Natural Language Processing (NLP), this dichotomy is particularly evident. Neural
models, such as transformers, have revolutionized NLP, enabling applications like machine
translation, text generation, and sentiment analysis. However, these models often lack the ability to
reason about language and its context in a human-like manner. Symbolic approaches, which are
rooted in linguistic structures and formal logic, can model complex relationships in language but
have difficulty scaling to handle the variability and ambiguity of natural language.
Neuro-Symbolic AI for NLP aims to bridge this gap by integrating neural models' data-driven pattern
recognition capabilities with symbolic reasoning's structured logic and explainability. This hybrid
approach enables systems to not only process and understand language but also to reason about it,
offering more interpretability, robustness, and the ability to generalize across different NLP tasks.
As a result, Neuro-Symbolic AI presents a promising pathway toward more advanced and human-
like NLP systems, opening up possibilities in areas such as explainable AI, improved language
understanding, and reasoning-based question answering.
1
WHY NEURO SYMBOLIC AI IN NLP
Definition : Neural networks excel at recognizing patterns in large volumes of text data,
enabling models to perform tasks like machine translation, text classification, and sentiment
analysis. However, these models often struggle with tasks requiring logical reasoning or
common sense understanding. Symbolic AI, in contrast, is based on rules and logic, which
makes it ideal for tasks that involve reasoning and decision-making. Neuro-Symbolic AI
leverages the strengths of both, combining neural networks' ability to understand patterns in
data with symbolic systems' reasoning capabilities.
One of the major criticisms of deep learning models in NLP is their black-box nature. Neural
models, while powerful, often do not provide clear insights into how they arrive at decisions or
predictions. By incorporating symbolic reasoning, Neuro-Symbolic AI systems become more
transparent and explainable, making it easier to understand and trust their outputs, especially in
critical applications like healthcare or legal NLP systems.
Natural language is inherently ambiguous and context-dependent. Neural models may struggle to
fully capture nuances, like multiple meanings of words or the relationships between different entities
in a sentence. Symbolic reasoning, which works with structured knowledge, can help models
disambiguate language and understand context better. For instance, it can employ formal logical
rules or knowledge graphs to make sense of complex sentence structures or ambiguous terms,
enhancing language comprehension
neural models can overfit to specific training data and may lack the ability to generalize to new,
unseen contexts. Symbolic AI, with its reliance on rules and logic, provides a more structured
2
approach that is inherently generalizable across different language tasks. By combining these
approaches, Neuro-Symbolic AI can produce systems that are not only good at specific tasks but are
also more robust and adaptable to a wide range of NLP applications.
Many advanced NLP tasks require more than just pattern recognition; they require deep
understanding and reasoning. Tasks like question answering, semantic parsing, and natural
language understanding often involve both learning from data and applying reasoning rules.
Neuro-Symbolic AI is particularly well-suited for such tasks, as it allows for the integration of
learned representations from neural models with logical, rule-based inferences from symbolic
systems.
With the growing need for ethical AI and AI accountability, Neuro-Symbolic AI helps address
concerns about the opacity of neural models in NLP. Explainable NLP models that combine
symbolic reasoning offer a way to track the decision-making process, helping users understand why
a system made a specific prediction or action. This is crucial for industries where trust and
fairness are non-negotiable.
Neural models often require vast amounts of labeled training data, which can be expensive and time-
consuming to generate. Symbolic AI, on the other hand, can work efficiently with predefined rules
3
and logical structures, reducing the reliance on large datasets. In Neuro-Symbolic AI, symbolic
components can help compensate for limited training data by providing structured guidance to the
neural components.
One of the ultimate goals of NLP is to create systems that can understand language like humans, not
just by recognizing patterns, but by reasoning and making sense of context and meaning. While
neural networks excel at mimicking certain aspects of human cognition (like identifying patterns in
language), they often fall short in logical reasoning, common sense understanding, and abstract
thinking—areas where humans naturally excel.
4
CORE COMPONENTS
The core components of Neuro-Symbolic AI for NLP involve the integration of neural models, which
excel in pattern recognition, with symbolic reasoning systems, which provide structured logic-based
reasoning. This hybrid approach helps in addressing the complexities of natural language. The main
components include:
Neural networks form the backbone of modern NLP applications. These models are capable of
learning from large datasets and capturing intricate patterns in natural language. Common neural
architectures in NLP include:
Recurrent Neural Networks (RNNs): Used for sequential data processing, like in language
modeling and machine translation.
Convolutional Neural Networks (CNNs): Applied for tasks like text classification, capturing local
dependencies in text.
Transformers: Revolutionized NLP with models like BERT, GPT, and T5, enabling tasks like text
generation, question answering, and translation through self-attention mechanisms.
Neural models are highly effective at handling ambiguous, noisy, or unstructured language data,
making them useful for tasks like machine translation, sentiment analysis, and named entity
recognition.
reasoning involves using predefined rules, logical expressions, and knowledge representations to
understand and process language. While symbolic methods are less common in modern NLP, they
excel in areas requiring logical inference, structure, and interpretability. Key components of
symbolic reasoning in NLP include:
Logic-Based Systems: Symbolic models often use formal logic to reason over natural language, such
as propositional or first-order logic.
Grammar Rules and Parsing: Traditional symbolic methods in NLP use grammars to parse and
understand sentence structures, such as syntactic parsing.
5
Knowledge Representation and Ontologies: Symbolic systems often utilize structured knowledge
bases or ontologies to represent and reason about entities, relationships, and facts in natural language.
Knowledge graphs are often used in Neuro-Symbolic AI to represent facts, relationships, and
ontological structures. These graphs can be used in NLP to enhance reasoning capabilities and
provide context beyond what neural networks alone can infer.
Entities and Relations: Knowledge graphs represent entities (nouns) and their relations (verbs or
prepositions), enabling systems to understand the context and meaning in a structured way.
Integration with Neural Models: Neural models can be trained to predict or embed entities and
relationships into vector spaces, while symbolic systems can query the graph for logical inferences
and knowledge retrieval.
The integration of neural networks and symbolic reasoning is at the heart of Neuro-Symbolic AI.
There are various methods to combine these two components:
Neural Models Guided by Symbolic Rules: Neural networks can be guided by symbolic rules or
logic constraints, ensuring outputs adhere to certain logical consistency, improving interpretability
and robustness.
Neural Networks with Logical Constraints: Techniques like Logical Neural Networks
incorporate logical rules directly into the learning process of neural networks, ensuring that model
predictions respect the symbolic rules.
Symbolic Reasoning over Neural Representations: After a neural model has processed text (e.g.,
identifying relevant features or entities), symbolic reasoning systems can perform logical operations
or infer higher-level insights based on the neural outputs.
The learning and inference mechanisms in Neuro-Symbolic AI combine the data-driven learning of
neural models with the reasoning capabilities of symbolic methods:
Neural Learning: Neural networks learn from large-scale text datasets, identifying patterns,
relationships, and semantic meanings that are difficult to encode explicitly.
6
Symbolic Inference: Symbolic systems, using logic or rule-based approaches, can perform
inference over the knowledge extracted by neural models, leading to more structured and
interpretable decisions.
of the key advantages of Neuro-Symbolic AI is its potential for greater transparency and
explainability:
Symbolic Reasoning for Explanations: Symbolic reasoning components can provide interpretable
explanations for decisions made by the neural components, such as why a particular prediction or
classification was made.
Rule-Based Reasoning: Symbolic rules can be used to explain the relationships and logic behind
language models’ outputs, providing users with more trust in AI systems.
A critical challenge in NLP is understanding and preserving the context of sentences and language.
Neuro-Symbolic AI components address this by:
Neural Embeddings for Context: Neural models like transformers capture contextual information
through word or sentence embeddings, which represent words in relation to their surrounding text.
Symbolic Representations for Logic: Symbolic systems can model complex logical relationships,
ensuring that language is processed with context and reasoning in mind, such as understanding
negation, causality, or temporal dependencies.
7
APPLICATIONS OF NEURO SYMBOLIC AI FOR NLP
8
Text-Based Reasoning and Decision Support Systems
Neuro-Symbolic AI is particularly valuable in decision support systems for domains such as law,
finance, and medicine. By combining neural models for data processing with symbolic reasoning
for applying rules or logic, these systems provide better recommendations and decision-making,
with clear justifications based on structured knowledge.
9
INTEGRATION OF NEURAL AND SYMBOLIC APPROACHES
Integrating neural and symbolic approaches in Neuro-Symbolic AI combines the strengths of both
paradigms to handle complex tasks in Natural Language Processing (NLP). Here are some common
methods for integration:
Hybrid Architectures
End-to-End Models: These models merge neural and symbolic components into a single architecture.
For instance, a neural network might perform pattern recognition and feature extraction from text,
while a symbolic reasoning system processes these features to apply rules or make logical inferences.
The outputs from both components are then combined to produce the final result.
Sequential Models: In this approach, neural and symbolic components operate sequentially. For
example, a neural network might first generate intermediate representations of the text, which are
then analyzed by a symbolic reasoning system for further processing and decision-making.
Neural-Symbolic Embeddings
Representation Learning: Neural networks can create embeddings representing symbolic knowledge,
such as logical rules or ontologies. These embeddings allow symbolic information to be integrated
into neural models, enabling them to use structured knowledge in their learning process.
Knowledge Encoding: Symbolic knowledge is converted into a format that neural networks can
process. For example, symbolic rules or knowledge graphs might be transformed into vector
representations that are fed into neural models, helping them utilize external knowledge during
training and inference.
Rule-Based Regularization: Symbolic rules can be applied as constraints or regularization terms during
neural model training. This ensures that the models adhere to certain logical rules or structures,
improving their performance on tasks requiring rule adherence.
Guided Learning: Symbolic reasoning can guide neural network learning by providing additional
signals or feedback based on logical rules. For example, symbolic systems might highlight errors or
inconsistencies in neural model outputs, helping to refine and enhance the model.
10
Modular Systems
Component Integration: In modular systems, neural and symbolic components are developed as
separate modules that interact with each other. For instance, a neural network module might handle
language understanding, while a symbolic module manages reasoning tasks. The outputs from both
modules are combined to achieve the final result.
Interfacing Layers: Interfacing layers facilitate communication between neural and symbolic modules,
ensuring seamless information flow and allowing both components to work together effectively
while maintaining their distinct functions.
Refinement of Outputs: After a neural network generates its output, symbolic reasoning can be used
to refine or validate the results. For example, neural models might produce candidate answers for a
question, and a symbolic reasoning system can verify or rank these answers based on logical rules
or external knowledge.
Validation and Correction: Symbolic systems can validate and correct neural model outputs by
applying logical checks or consistency rules, ensuring that the results are coherent and adhere to
expected norms or constraints.
Interactive Learning
Active Learning: Symbolic reasoning can guide active learning processes, where the neural model
learns iteratively from feedback provided by the symbolic system. For instance, the symbolic
component might identify areas where the neural model is weak and suggest specific training
examples to address these weaknesses.
Human-in-the-Loop: Combining neural and symbolic approaches allows for interactive learning
scenarios where human experts provide input and corrections based on symbolic reasoning. This
improves the system’s performance and adaptability to specific tasks.
Knowledge Transfer: Symbolic knowledge can be used to initialize or fine-tune neural models,
leveraging existing structured information to enhance model performance. For example, pre-trained
11
neural models can be fine-tuned with symbolic rules or knowledge to adapt to specific domains or
tasks.
Domain Adaptation: Symbolic reasoning aids in adapting neural models to new domains by providing
domain-specific rules or knowledge that guide the learning process and improve performance in
specialized areas.
Rule-Based Explanations: Symbolic systems can generate explanations based on predefined rules and
logic, complementing the neural model’s outputs and providing clear justifications for the results.
12
INFORMATION EXTRACTION AND KNOWLEDGE REPRESENTATION
Information extraction (IE) is the process of identifying and extracting structured information from
unstructured text. This process involves several key steps:
Entity Recognition: Identifying and classifying named entities such as people, organizations, locations,
dates, and other relevant terms within the text. Neural models, such as named entity recognition
(NER) systems, can detect these entities based on learned patterns from large datasets.
Relationship Extraction: Determining and extracting relationships between identified entities. For
example, in a sentence like "John works at Acme Corp," the relationship between "John" and "Acme
Corp" is an employment relationship. This often involves analyzing the context around entities to
identify how they are related.
Event Extraction: Identifying events and their participants from the text. An event might include an
action, a location, and participants. For instance, in "The conference will be held in New York next
month," the event is the conference, with "New York" as the location and "next month" as the time.
Fact Extraction: Extracting factual information such as data points, statistics, or specific pieces of
knowledge. This often involves detecting and recording specific details mentioned in the text.
Neural approaches, including deep learning models and transformers, have made significant advances
in information extraction by learning patterns and structures from large amounts of text data.
However, integrating symbolic reasoning can further refine the extraction process by applying
predefined rules and knowledge to ensure accuracy and consistency.
Knowledge Representation
13
Knowledge Graphs: Knowledge graphs are a form of ontology that represents information as a network
of entities and relationships. Each entity is a node, and each relationship is an edge connecting nodes.
Knowledge graphs facilitate complex queries and reasoning by organizing information in a
structured format.
Semantic Networks: Semantic networks represent knowledge in terms of nodes (concepts) and edges
(relationships) between them. These networks can be used to model complex relationships and
support reasoning about the connections between different concepts.
Rule-Based Systems: Rule-based systems use a set of predefined rules to represent knowledge and
perform logical reasoning. For example, a rule-based system might use if-then rules to infer new
facts from existing ones or to apply logical constraints.
Frame Representation: Frame representation involves organizing knowledge into structures called
frames, which capture stereotypical situations or concepts. Each frame contains slots for different
attributes and values, providing a way to represent and reason about complex information.
Neural networks can enhance knowledge representation by generating embeddings that capture semantic
meaning and context from text. These embeddings can be used to populate and enrich knowledge
graphs or ontologies. Symbolic reasoning, on the other hand, provides a way to apply logical rules
to the structured knowledge, ensuring consistency and supporting complex inferences.
Integrating neural approaches with symbolic knowledge representation can lead to more robust
information extraction and richer knowledge models. Neural networks provide the capability to learn
from vast amounts of text data, while symbolic systems offer the structure and reasoning capabilities
needed to apply and interpret that knowledge effectively.
14
EXPLAINABILITY IN NLP
Explainability in natural language processing (NLP) refers to the ability to understand and interpret how
and why an NLP model produces a particular output or decision. This is crucial for building trust in
AI systems, especially in applications where decisions impact human lives or where regulatory
compliance is required. Here are key aspects of explainability in NLP:
Model Interpretability
Feature Importance: Understanding which features or inputs significantly influence the model's
predictions. Techniques such as feature importance scores or attention maps help highlight which
parts of the input text are most relevant for a given prediction.
Visualization: Using visual tools to display how a model processes input data. For example, attention
maps in transformer models show which words in a sentence the model focuses on when making
predictions, offering insights into its decision-making process.
Model Design: Knowing how the model is structured, including its layers, activation functions, and
learning mechanisms. This transparency helps in understanding the model's overall behavior and
how it handles different types of inputs.
Algorithmic Explanation: Providing insights into the algorithms used by the model, such as how it
applies rules or learns from data, contributes to a deeper understanding of its operations.
Explanation Methods
Local Explanations: Techniques that explain individual predictions. For example, Local Interpretable
Model-agnostic Explanations (LIME) or Shapley values can be used to interpret the contribution of
each feature to a specific prediction, providing localized insights into why a particular output was
generated.
Global Explanations: Methods that provide an overview of the model's behavior across all predictions.
For instance, examining the distribution of predictions or the overall impact of different features
helps understand general trends and model behavior.
Rule-Based Explanations
15
Logical Rules: In systems incorporating symbolic reasoning, the explanation can be based on logical
rules or constraints applied to the input data. This approach provides clear justifications based on
predefined rules, making the reasoning process more transparent and understandable.
Decision Trees: Using decision trees or rule-based models can provide intuitive explanations for
predictions by showing how decisions are made based on specific criteria or rules.
Human-AI Interaction
Interactive Feedback: Allowing users to interact with the model and provide feedback can help explain
and refine the model's outputs. For example, users can ask follow-up questions to clarify why a
certain prediction was made or provide corrections that improve the model’s accuracy.
Explanation Interfaces: Creating interfaces that present explanations in an accessible and user-friendly
manner. These interfaces might include natural language explanations or visual summaries that help
users understand the model’s decisions.
Trust Building: Explainability helps build trust in NLP systems by providing users with insights into
how decisions are made. This is particularly important in high-stakes applications such as healthcare,
finance, or legal contexts.
Accountability: Providing explanations ensures that models can be held accountable for their decisions.
If a model produces biased or incorrect results, explanations can help identify the sources of errors
and guide improvements.
Regulatory Requirements: Many industries have regulations that require AI systems to be explainable.
For example, the General Data Protection Regulation (GDPR) in the European Union includes
provisions for the right to explanation, which mandates that individuals be informed about the logic
behind automated decisions affecting them.
Complexity vs. Explainability: There is often a trade-off between model complexity and explainability.
More complex models, such as deep neural networks, tend to be less interpretable, while simpler
16
models, such as decision trees, are easier to explain but may not capture complex patterns as
effectively.
Balancing Accuracy and Transparency: Ensuring that explanations do not compromise the accuracy
of the model’s predictions. Striking the right balance between providing understandable explanations
and maintaining high performance is a key challenge in developing explainable NLP systems.
In summary, explainability in NLP involves making the workings and decisions of models
understandable to users. This encompasses various techniques and approaches, from visualizing
model behavior to using rule-based explanations and interactive feedback. Explainability is essential
for building trust, ensuring accountability, and meeting regulatory requirements in AI systems.
17
CHALLENGES AND LIMITATIONS
Complexity of Models
Understanding Deep Models: Deep learning models, such as transformers and large neural networks,
are highly complex with numerous layers and parameters. This complexity makes it difficult to
interpret how they arrive at specific predictions or decisions.
Opaque Decision-Making: As models become more intricate, their decision-making processes become
less transparent, making it challenging to provide clear explanations for their outputs.
Model Performance vs. Interpretability: Highly accurate models like deep neural networks often
sacrifice interpretability. Simplistic models, such as decision trees, are easier to explain but might
not achieve the same level of performance on complex tasks.
Balancing Act: Striking a balance between a model's accuracy and its explainability is a persistent
challenge. Improving one aspect can sometimes lead to compromises in the other.
Lack of Standardization
Inconsistent Methods: There is no universally accepted method for explaining NLP models. Various
techniques, such as LIME, SHAP, or attention mechanisms, provide different types of explanations,
and their effectiveness can vary based on the model and application.
Evaluation of Explanations: Evaluating the quality and usefulness of explanations is difficult. There are
no standardized metrics for measuring how well an explanation clarifies a model's behavior or aids
users in understanding its decisions.
Context Sensitivity
Dependence on Input Context: Explanations may vary depending on the context of the input text. For
example, the same model might provide different explanations for similar predictions depending on
subtle contextual differences, complicating the task of providing consistent explanations.
Dynamic Environments: In real-world applications, the context and data can change rapidly, making it
challenging to maintain consistent and accurate explanations over time.
18
Superficial Explanations: Some explanation methods may provide only superficial insights into a
model's behavior, such as highlighting specific words or features without addressing the underlying
reasoning process.
Lack of Depth: Explanations might not always capture the full complexity of the model's decision-
making process, leaving out important details that could impact understanding.
User Understanding
Technical Complexity: Explanations may be difficult for non-experts to understand. Technical jargon or
complex visualizations might not be accessible to all users, hindering the effectiveness of the
explanation.
Tailoring Explanations: Providing explanations that are both accurate and comprehensible for a diverse
audience is challenging. Different users may require different levels of detail or types of explanations
based on their expertise and needs.
Inherent Biases: If a model is biased, explanations might also reflect or obscure these biases, rather than
providing a neutral understanding of the model’s behavior.
Fairness in Explanations: Ensuring that explanations are fair and do not inadvertently perpetuate biases
or reinforce negative stereotypes is an ongoing challenge.
Data Dependence
Training Data Influence: Explanations are influenced by the training data used to develop the model. If
the data is biased or incomplete, the explanations might also reflect these limitations, affecting their
accuracy and fairness.
Generalization Issues: Models might perform well on specific datasets but provide less reliable
explanations for different or unseen data, limiting their generalizability.
Computational Costs
Resource Intensity: Some explanation techniques, such as generating attention maps or running
additional models for local explanations, can be computationally expensive and resource-intensive.
Scalability: Scaling explainability methods to work efficiently with large datasets or in real-time
applications can be challenging, impacting their practical usability.
19
In summary, the challenges and limitations of explainability in NLP include the complexity of models,
the trade-off between accuracy and interpretability, lack of standardization, context sensitivity, and
user understanding. Additionally, issues related to bias, data dependence, and computational costs
further complicate the task of providing clear and useful explanations. Addressing these challenges
requires ongoing research and development to create more transparent, fair, and accessible NLP
systems.
20
CONCLUSION
In conclusion, explainability in natural language processing (NLP) is crucial for building trust, ensuring
transparency, and meeting regulatory requirements in AI systems. As NLP models become
increasingly complex, achieving a balance between accuracy and interpretability presents significant
challenges. While advanced models such as deep neural networks offer impressive performance,
they often lack the transparency needed for users to understand their decisions fully.
Efforts to enhance explainability involve various techniques, including feature importance analysis,
visualization methods, and rule-based explanations. Each method has its strengths and limitations,
and there is no one-size-fits-all solution. The field continues to evolve, with ongoing research
focused on developing more effective and user-friendly explanation methods.
Addressing the challenges of explainability—such as the trade-off between model performance and
interpretability, the lack of standardized methods, and the need for context-sensitive explanations—
requires a multi-faceted approach. Improvements in these areas will contribute to creating more
transparent, fair, and accountable NLP systems.
Ultimately, enhancing explainability in NLP will lead to more robust and trustworthy AI systems,
fostering greater user confidence and facilitating better decision-making in a wide range of
applications. As the technology advances, it is essential to keep refining explanation techniques to
address the evolving needs of users and stakeholders in the AI ecosystem.
21
REFERENCES
Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). "Why should I trust you? Explaining the predictions
of any classifier." In Proceedings of the 22nd ACM SIGKDD International Conference on
Knowledge Discovery and Data Mining (KDD '16).
Sundararajan, M., Taly, A., & Yan, Q. (2017). "Axiomatic attribution for deep networks." In
Proceedings of the 34th International Conference on Machine Learning (ICML '17).
Chen, J., Song, L., Wainwright, M. J., & Jordan, M. I. (2018). "Learning to explain: An
information-theoretic perspective on model interpretation." In Proceedings of the 35th International
Conference on Machine Learning (ICML '18).
Caruana, R., Gehrke, J., Koch, P., Nair, R., & Ray, P. (2015). "Intelligible models for healthcare:
Predicting pneumonia risk and hospital 30-day readmission." In Proceedings of the 21th ACM
SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '15).
URL: https://dl.acm.org/doi/10.1145/2783258.2783310
Doshi-Velez, F., & Kim, B. (2017). "Towards a rigorous science of interpretable machine learning."
In Proceedings of the 2017 ICML Workshop on Human Interpretability in Machine Learning (WHI
'17).
Miller, T. (2019). "Explanation in artificial intelligence: Insights from the social sciences." In
Proceedings of the 2019 Conference on Fairness, Accountability, and Transparency (FAccT '19).
Zhang, S., & Chen, J. (2021). "A survey on explainable artificial intelligence (XAI)." In IEEE
Transactions on Neural Networks and Learning Systems.
Kim, B., & Kim, H. (2020). "Interpretability of machine learning-based models in health care." In
Journal of Biomedical Informatics.
22
23