Seunghan Han
Seunghan Han
Seunghan Han
genehmigten Dissertation.
"He who walks with the wise grows wise." Proverbs 13:20. First and foremost I would like
to thank my sincere advisors, Professor Walter Stechele and Dr. Andreas Hutter for the
guidance, support, and inspiration during my pursuit of the Ph.D. degree at TUM. I owe
and Professor Ebrould Izquierdo for reviewing and offering me advice during their busy
time. I also have to thank Dr. Ramesh Visvanathan and Dr. Vinay Shet for the multiple
SCR, I could get better insight on computer vision related researches. I would like to thank
Siemens and German Gorvernment because the work presented here was partially funded
TUM, including Colin Estermann, Andreas Laika, Tobias Schwarze, Jan Neumann, Vasu
Gero Bäse and Bonjung Koo for the many times of invaluable discussions on research
My sincere friends Daewoo Kim, Sungjin Kwon, Hoochang Shin, Changwook Lee
and Kiwon Park also deserve special thanks. Daewoo Kim was my continuous source of
joy for the discussions and talks not only about technical trends but also about life and
philosophy. Sungjin Kwon especially showed his support offering some of his technical
source codes on image processing. As a Ph. D candidate, Hoochang Shin and Changwook
III
Lee showed their interest on my work and gave me valuable comments. The CCTV video
I would like to thank my family and relatives for their love and encouragement.
source of support. There are no words to describe my mom’s generosity, love and kindness
In particular, I want to thank my dear girl friend Seung-hee. She has made my Ph.D.
career in munich so fruitful and enjoyable not only through her emotional support but also
through her scientific view as a medical researcher. Especially, her attitude towards life
pursuit of the Ph.D. degree, I realized how God made humans amazingly and miraculously
intelligent. Without his love, I know that I could not even finish this work. Best thanks to
God, the name of Love. "The fear of the LORD is the beginning of wisdom, and knowledge
IV
Abstract
Most of the automated and intelligent visual surveillance systems, so far, have focused
usual approach is building specific analytic algorithms for well-defined and fixed domains,
(e.g. pedestrian detection and tracking, left baggage detection, intrusion detection, etc.),
Thereby, the semantic space of the approach is bound to the results and capacity of used
algorithms. When this kind of system is used in high-level forensic, such an approach is
inherently too limited. This is due to the fact that modeling all situations at develop-
ment time is very difficult. The difficulty arises not only from the limited vision analytic
power on signal level vision information processing but also from the ambiguity inherent
interpretation by collecting visual evidences and combining such evidences with contextual
knowledge. Such a human activity includes derivation, modification and retraction of dif-
ferent semantic conclusions upon arrival of new information. Thus, it can be regarded as
non-monotonic reasoning
Bearing such a human ability in mind, the presented work focuses on designing a sys-
tematic approach and methodical support for a more intelligent forensic visual surveillance
dences and focus more on the way of manipulating such evidences to derive more complex
semantics of interest. We approach the manipulation from the epistemological stance, that
is a theory aimed at targeting notions such as knowledge, belief, awareness and proposi-
tional attitudes, by means of logical and semantical tools. First, we propose the use of logic
programming for the basis of knowledge representation and reasoning. Second, we propose
V
the use of subjective logic to capture uncertain epistemic belief about knowledge. Third,
upon the proposed subjective logic extended logic programming framework, we present
an inference scheme for so-called default reasoning, that can draw plausible reasoning re-
model uncertain and ambiguous contextual rules, based on so-called reputation operator
not allow bidirectional interpretation of a rule. Based on so-called deduction and abduc-
tion operators in subjective logic we show that bidirectional reasoning can be done in our
framework. Finally, by the use of abductive logic programming with subjective logic, we
demonstrate a reasoning framework that can answer most feasible hypotheses to the best
fore-mentioned aspects are presented with case studies from typical public area scenes with
VI
Contents
1 Introduction 1
i
Contents
3 Preliminaries 37
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
ii
Contents
4.6 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
5.7 Multivalued Default Logic, Square Bilattice and Default Opinion . . . . . . 126
iii
Contents
5.9.1 Which way to go? Subjective Logic vs. L-fuzzy Logics . . . . . . . . 144
iv
Contents
9 Conclusion 223
v
Contents
Bibliography 239
vi
Chapter 1
1 Introduction
Epistemology
In the security sector of industry, visual surveillance is one of the oldest and most widespread
technologies as cameras are quite informative and relatively inexpensive sensors. Since the
hundreds of cameras, etc. Consequently, as it is getting digitized, we also get huge amount
of video information that can not be processed only by human security agents. Therefore,
success. However, this is an extremely difficult task for visual surveillance systems1 . To
address this problem, there has been active research focus and paradigm shift for the last
unexpected and potentially dangerous situations for immediate intervention, most of the
automated and intelligent visual surveillance systems, so far, have focused on real-time
1
Contents
Figure 1.1: Research Spectrum of Intelligent Semantic Analysis in Visual Surveillance and
Paradigm Shift.
building specific analytic algorithms for well-defined and fixed domains (e.g. pedestrian
detection and tracking, left baggage detection, intrusion detection, etc). Such analytics
usually model activities or events as a sequence of a number of ‘states’ that are related
to visual features (such ‘state model’ based approaches are also referred to as ‘intensional
approach’ [130]). Therefore, in this paradigm the focus is more on ‘automated perception’
of specific events as depicted in Figure 1.1 - ‘2nd Paradigm’. Consequently, due to the
difficulty of modeling all situations at development time, the semantic space is inherently
Following the September 2001 attacks on the United States [134] and then those in
London in 2005 [10], it was proven that the primary objective to prevent all potentially
possible threats is not always achievable [148, 74] even in cities where a huge number of
surveillance cameras was deployed (London is the city the most often cited for the number
of cameras deployed in its streets. [74]). Naturally, forensic use of visual surveillance in
an intelligent way is gaining increasing attention, not only for real-time observation but
2
Chapter 1
Contents
also for post incident investigation. Such investigation usually deals with propositional
such higher level semantics is not a trivial task. The difficulty arises not only from the
ambiguity inherent in visual information and limited vision analytic power but also from
the ‘complex plot’ [148] implied in video scenes in case of planned or intended events.
combining the evidences with contextual knowledge. This can be regarded as ‘cognitive’
human activity that deal with fusion of ‘propositional context knowledge’ and ‘belief ’ on it.
logical and semantical tools [137, 138]. By this aspect, we advocate that high level semantic
Although it is undergoing the paradigm shift, relatively little work from the ‘auto-
mated epistemic reasoning’ stance has been done. This dissertation is intended to address
this void in semantic analysis research. Thus, the main objectives of this dissertation is as
follows.
• Designing a systematic approach for reuse of metadata acquired from vision analytics.
• Introducing a methodical support for epistemic reasoning under incomplete and uncer-
• Studying the feasibility of the proposed systematic approach and methodical support for
In visual surveillance systems, metadata acquired from vision analytic modules can
2
Human ‘cognition’ consists of many aspects such as psychology, neuroscience, logics, linguistics, sociology,
etc., and ‘cognitive science’ is the study on these aspects. While ‘artificial intelligence’ aims to model
some of these aspects for a machine. Among these aspects, ‘epistemology’ addresses the question
‘How do we know what we know?’ and deals with the means of production of knowledge, as well as
skepticism about different knowledge claims. In artificial intelligence, therefore, it is closely related to
logics, automated judgement of belief, etc. For more detailed overview, refer to [116, 138, 137].
3
Contents
be regarded as symbolized visual evidences obtained by artificial vision sensors from the
2nd paradigm view. Then, such evidential metadata should be epistemically fused by con-
a set of semantic rules, constraints and relations, i.e., in particular, this is the part that
has a large degree of overlap between how humans describe what constitutes a situational
concept and how it is defined within contextual knowledge modeling formalisms. In other
words, recognizing certain high level contextual semantic concept as it occurs becomes a
problem of ‘explaining’ the observation using the available semantic knowledge. Therefore,
the higher level semantic models should contain modeling approaches that do not just
aim to define the entire ‘state’ space of the event domain as in ‘state model (intensional)’
approaches. Instead, we advocate that it should rather enable human-like ‘epistemic rea-
soning’ that can offer a formal way how we derive and guarantee the target semantics of
• What are the semantic queries that can be answered based on epistemic reasoning under
• How can one make the ‘sensing (automated perception)’ result of available vision analytic
modules reusable ?
• Where does the semantic knowledge come from ? How can one represent semantic know-
ledge in a machine understandable way ? What is the proper formalism for representing
• How humans assess propositional assumptions and contextual concepts along the given
evidences and contextual knowledge ? For example, humans tend to assess a proposition
in a vague manner such as ‘strongly certain’, ‘less certain’, ‘seem to be true but quite not
4
Chapter 1
Contents
• Can one guarantee that a given contextual knowledge is always sound and complete ?
What are the epistemic aspects that can arise under such a circumstance and how can
• The results from vision analytic modules usually come with uncertainty, and their per-
formance may vary according to its internal analytic mechanisms and training data.
How can one take into account such uncertainty at the time of reuse ?
• What are the practical and pragmatic building blocks given currently available software
tematic approach and methodical support for a more intelligent higher level reasoning
capability.
This subsection details characteristics of visual surveillance and derives challenges and
requirements in terms of forensic sense of high level semantic queries (refer to [49, 74] for
In the sense that ‘automated epistemic reasoning’ rely on currently available vision analytic
ized as follows.
1. Environment : Distributed
2. Processable Semantics :
5
Contents
3. Types of data :
- Alarms, objects and events representable in symbolic form (e.g. text format, etc)
6. Semantic complexity of query : Very high and vary from specific to abstract and vague
8. Performance metrics :
derive more concrete requirements on the components and the system itself for complex
queries. We analyze real events, e.g. in which the authorites released CCTV footage
after closing investigation on the subway bombing attacks in London on 7th july 2005. A
question roadmap [10] shows the questions dealt by London Police such as ‘Who is the
suspect?’, ‘Was it done by a single person or by a group?’, ‘Why did they do it?’, ‘How
did they do it?’, etc. To be machine supportable, it is important to take into account the
semantic granularity of the query. In the example, the ‘why’ or ‘how’ questions seem much
harder than ‘who’ questions because the more knowledge about the situation is required.
compositional, that can be tied with other atomic semantic meanings with visual cues and
situational context. Figure 1.2 (left) shows steps gathering situational knowledge and how
the acquired knowledge influences the iterative reasoning especially for forensic sense of
search and retrieval. Figure 1.2 (right) shows one sample of the released footage. Although
6
Chapter 1
Contents
Figure 1.2: Conceptual Workflow of Forensic Reasoning in Case of Video Retrieval (left)
and a Released Retrieved-Footage (right) [10].
this sample seems very ordinary, the people in the scene were regarded as suspects of a
terror when considered in the situational context. Therefore, interpreting the semantic
For a more concrete discussion, we assume a distributed camera network and an unexpected
incident not directly captured by a camera due to the sparseness of the network and lack
of sufficient vision analytic power. Assuming some clues, let us consider following queries.
Conjunctive queries that can be seen as existence of some of specific atomic metadata
Compositional queries having relational semantics that may require more contextual in-
formation and that may require iterative and active collection of contextual information.
(e.g. group of, passing by, carrying, talking to, witness of, theft of, etc).
7
Contents
- Q4 ‘scenes in which a person is carrying an oil box around the place of the reported
incident’
- Q5 ‘scenes one of selected person, among retrieved after Q4, is talking to somebody’
- Q10 ‘why did the woman detected in a robby suddenly went to outside?’
In the following, by considering above query scenarios, the most important requirements
towards target high level semantic reasoning capability and systems are described.
imize the potential utilization for later query and reasoning processing. As shown in
Section 1.2.1 - 2, manual annotation could be also used to denote the concept ‘telephone
booth’ in Q3. Similarly, results from automatic human detection algorithm could be
exploited to serve as atomic ingredient for processing more complex concepts such as
‘group of people’, ‘witness’, ‘suspect’ and ‘theft’ in Q1-Q9. This should include low-level
like a suited metadata representation model is required to archive the processing result.
8
Chapter 1
Contents
time, the system requires an efficient way to describe contextual knowledge about scenes.
Unlike in the deterministic ‘state-model (intensional)’ based approach [130], here the
reasoning target is not given at development time. Furthermore, a given problem can
not be solved directly by a known set of steps (see Q6-Q8, Q10-Q12) and the input data
is ‘incomplete’ because the available results generated by the analytics can not be fully
free’ approach is also not possible. Thus, we rather need a more flexible ‘extensional’
semantic inference mechanism that allows ‘modularity’ of each knowledge segments [130].
diate metadata comes with ‘uncertainty’ (e.g. trust worthiness of results and errors in
this will influence the acceptable accuracy and robustness for a semantically complex
queries. Thus, a mechanism is required for dealing with the varying quality to the re-
trieved results, i.e. to cope with the ‘uncertain’ nature of intermediate results captured
antee the ‘soundness’ and ‘completeness’ of given contextual knowledge. There are
‘uncertainty’ in knowledge itself. For example, there can be often situational ‘contra-
segments (see Q7-Q9). Therefore, we need a mean to assess ‘epistemic belief ’ to attach
be revised upon arrival of new evidential data or contextual knowledge. Such capability
9
Contents
• Epistemic Meta Reasoning and Abduction : in the case of highly abstract and
reasoning [45]) power that can even able to reason about the query itself in an iterative
manner. The system should be able to set ‘epistemic hypothesis’ by itself and assess
is known as ‘abduction’ [150] (note that, especially, more complicated handling of this
aspect is closely related to ‘planning’ in the view of artificial intelligent). Therefore, the
The requirements for ‘belief revision’, ‘default reasoning’, ‘meta-reasoning’ and ‘ab-
duction’ can be categorized under the term of ‘non-monotonic reasoning’ [35]. Considering
In this section, we present our approach to the requirements designated in the previous
section.
First, one of the most important aspects throughout above derived requirements is
ory [93] that is a relatively new branch of probabilistic logic which comes with a rich set
like epistemic belief in a model called ‘subjective opinion’. A subjective opinion not only
allows to describe one’s belief based on degree of truth or falsity but also allows explicit
representation of ignorance about the degree of truth. This aspect is inherited from Demp-
ster Shafer’s belief theory [149]. But it also provides a mathematical mapping between the
10
Chapter 1
Contents
opinion representation and 2nd order Bayesian especially, via a Beta function representa-
tion [20]. The intuition behind is that the uncertainty representation of subjective logic
covers traditional probability so that we can directly encode traditional bayesian based un-
certainty calculation results coming from vision analytic modules (in the sense that most
of ‘state model’ based approaches rely on Bayesian approach throughout model learning
and recognition phase). This aspect matches to the requirements of uncertainty attach-
intuitive and human like ‘epistemic status’ can be represented. This aspect matches to the
Second, for the semantic knowledge modeling and representation, we exploit logic
programing approach. We do not believe that we can derive a ‘whole-model’ that can cover
most of semantic situation. We also do not believe that a ‘model-free’ approach is possible
data is again extremely difficult task. Therefore, our choice is to bestow more degree of
freedom on modeling semantic knowledge. To achieve this purpose, the most important
aspect is that the system need to guarantee a ‘modularity’ so that any domain engineers can
just concentrate on their partial knowledge segment without considering whole knowledge.
Another aspect is that the knowledge representation formalism should provide a mean
also seen in a logic sense, in that a pair of nodes connected with edge can be seen as
‘if node1 then node2’ manner. However, this level of conditional description remains in
‘propositional logic’ that only offers one relation ‘influence’. Unlike propositional logic,
logic programming usually offers predicate logic that can build various relations. Although
there are other technologies for knowledge representation such as ‘entity-relation diagram’,
‘semantic network’, ‘ontology and semantic web’ they tend to rely on schematic approach
that more fit to data archiving aspect. Therefore, we adopt semantic web and database
11
Contents
introduce a principled reasoning scheme especially for ‘default reasoning’, ‘vague proposi-
tional rule’ modeling and ‘abductive explanation of observation’. The reasoning mechanism
is realized based on logic programming extended with subjective logic based operations on
In this way, we give more flexibility and on demand way of using both ‘intensional’ and
handling ‘contradictory information’ so that it can derive plausible reasoning result. For
the vague propositional rule definition, we introduce subjective reputation function for
we take both ‘deductive’ and ‘abductive’ logic programming. Especially, we take both
We now summarize the key contributions of this dissertation tackling the challenges of
• The idea to combine logic programming and subjective logic theory for flexible knowledge
• The study on system architecture for reuse of vision analytic metadata comparing tra-
• The design of methodical support for contradictory information handling using subjective
logic theory and its comprehensive comparison with L-Fuzzy set based approaches.
• The idea of using reputation concept for modeling vague propositional knowledge.
• The idea of using deduction and abduction operators of subjective logic for enabling
12
Chapter 1
Contents
• The idea of layering abductive and deductive logic programming for logical abduction
In the following Chapter 2, we give a brief historical review of epistemology and logic
both in terms of philosophical view and artificial intelligence. Then we review relevant
This will cover subjective logic and logic programming. Except the fundamentals explained
in the preliminaries (Chapter 3), the rest of chapters are organized to be self-contained.
subjective logic. We discuss the software components, data processing pipeline and system
extensional system especially in terms of so-called default reasoning. We present our ap-
subjective opinion space with square bilattice. We then models a default reasoning scheme
adopting the idea exploited in bilattice system. There, we also compare our approach to
L-fuzzy set based approaches. We then presents a case study on handling contradictory
on the comparison between proposed approach and L-fuzzy set based approaches such as
square bilattice.
13
Contents
and handling. There, we present the specific idea on evaluating a proposition with a
reputation function of subjective logic. We also describe a case study on reasoning under
vague propositional rules. We conclude the chapter with a discussion of the experimental
Chapter 7 describes our work on hybrid knowledge modeling and reasoning. First,
Then, we present the idea how to enable intensional handling of conditional rules in an
extensional framework using subjective logic. Finally, this chapter presents, in detail, the
In Chapter 8, we present our framework for logical abduction based scene explanation
under collected observations. There, we present the idea of layering deduction on query
itself and abduction to get final answer. This chapter also shows how the final answers
can be evaluated in terms of default reasoning. Finally, we conclude the chapter with
In Chapter 9, we present the conclusions of our work with discussions and summary
14
Chapter 2
2 Prior Art and Related Work
In this chapter, we will review related work ranging from a philosophical background of
follows with reviews of prior art ranging from a semantic multimedia retrieval to automatic
Epistemology is the branch of philosophy concerned with the nature and scope of knowledge
[11, 174]. In the view of philosophical history, since Aristotle, Platon to Descartes and Kant,
the majority of philosophy was metaphysics to explain the fundamental nature of being
and the world [105]. Since Aristole, epistemology was dealt as a sub topic of metaphysics
in the focus on structuring existence of entities in the world. Especially, ‘ontology’ received
much interests. Ontology is the theory of objects in terms of the criteria which allow
one to distinguish between different types of objects and the relations, dependencies, and
properties through which they may be described. Since Descartes, epistemology has paid
more attention on how humans think and how humans can sure their belief on knowledge.
Addressing this problem, we face with following fundamental question, "Is the limitation
of one’s cognition bounded via sensory experience ? or is it bounded via innated humans
rationality ?". Through 17 to 18 centuries, we bestowed the name ‘Empiricism’ on the first
category and ‘Rationalism’ on the other category. These two major trends are compromised
by Immanuel Kant in his remarkable work "Die Kritik der reinen Vernunft". Kant thought
humans do not know as knowledge exists but believe a certain knowledge as humans know.
15
Contents
Therefore, to him truth is something that we need to find out of structure of human
thinking [107]. This idea is depicted in Figure 2.1. In terms of rationalism, he introduces
‘a priori innate human rationality’ that serves categories on thinking as posed follows:
• Quantity (Quantität) :
• Quality (Qualität) :
• Relation (Relation) :
• Modality (Modalität) :
16
Contents
With the help of such priory categories, he believed that humans can fuse their
empirical sensing ability within these categories to extend their knowledge. This aspect is
Chapter 2
understood as ‘judgement’ and ‘propositional rules’ are the representation of judgement.
• Quantity (Quantität) :
• Quality (Qualität) :
• Relation (Relation) :
• Modality (Modalität) :
Humans in everyday life face with various situations and events. Then humans
think various ways to solve, analyze the facing situations and events. Of course the ways
falls into above categories. Categorical judgement (Kategorische Urteil), among others, is
the most basic one and every 12 types can be transformed into the categorical judgement
except ‘Hypothetical’ and ‘Disjunctive’. When the categorical judgment is considered with
17
Contents
‘universal’, ‘particular’, ‘affirmative’ and ‘negation’, following four types of basic judgement
is possible.
already known judgements are called ‘premises’. In case there is only one ‘premise’ we
call it ‘direct inference’ and in case of handling ‘multiple premises’, we call it ‘indirect
conclusion does follow necessarily from the premises, i.e., if the conclusion must be true
provided that the premises are true. A deductive argument is sound if it is valid and
its premises are true. Deductive arguments are valid or invalid, sound or unsound, but
are never false nor true. Deductive reasoning is a method of gaining knowledge. An
Therefore, Socrates is mortal. The first premise states that all objects classified as ’men’
have the attribute ’mortal’. The second premise states that ’Socrates’ is classified as a
man - a member of the set ’men’. The conclusion states that ’Socrates’ must be mortal
some degree of support (inductive probability) for the conclusion but do not entail it;
18
Contents
that is, they suggest truth but do not ensure it. Induction is employed, for example,
in the following argument: 1) Every life form we know of depends on liquid water to
Chapter 2
exist. 2.) All life depends on liquid water to exist. Inductive reasoning allows for the
possibility that the conclusion is false, even where all of the premises are true. For
example: 1) All of the swans we have seen are white, 2) All swans are white. Through
many dictionaries inductive inference are also defined as reasoning that derives general
Charles Sanders Peirce as "guessing" [131]. The term refers to the process of arriving
that a is sufficient (or nearly sufficient), but not necessary, for b. For example, the
lawn is wet. But if it rained last night, then it would be unsurprising that the lawn
is wet. Therefore, by abductive reasoning, it rained last night. (But note that Peirce
did not remain convinced that a single logical form covers all abduction.) Peirce argues
that good abductive reasoning from P to Q involves not simply a determination that,
e.g., Q is sufficient for P, but also that Q is among the most economical explanations
for P. Simplification and economy are what call for the ’leap’ of abduction. There has
been renewed interest in the subject of abduction in the fields of computer science and
In the context of traditional computer vision research on semantic analysis, the ma-
jority of the work focusses on ‘concept’ level as a result of automated perception. Although
the importance of judgements and inference, little attempts have been paid to ‘automated
judgements and inference’. In this dissertation, we especially pay more attention to the
‘automated inference’ in terms of ‘deduction’ and ‘abduction’. To cope with the quality
property of the judgement, we also pay attention for the type of ‘Particular (besondere
19
Contents
Urteil)’ judgement.
ontology’ that tries to encode the relational structure of concepts which one can use to
describe and reason about aspects of the world. Inheriting this tradition, today, formal logic
also focusses on artificially structured languages and its syntax or semantics. Traditional
knowledge representation models developed in the field of artificial intelligence are formal
logics [146], semantic networks [141, 84], frames [118, 84] , Description Logic [24, 2] and
Logics and semantic networks are widely accepted models for effective knowledge represen-
tation. Logics aim at emulating the laws of thought by providing a mechanism to represent
statements about the world - the representation language - and a set of rules to deduce new
statements from previous ones - the proof theory. The representation language is defined
by its syntax and semantics, which specify the structure and the meaning of the statements,
respectively. Different logics make different assumptions about what exists in the world
(e.g. facts) and on the beliefs about the statements. The most basic logic is Propositional
Logic [146]. Propositional logic declaratively deals with pieces of syntax correspond to
can be seen as a propositional logic in the sense that a conditional probability P (B|A) can
be considered as ‘If B then A’. Fuzzy logic [176] also deals facts that comes with degree of
truth. However, propositional logic has very limited expressive power unlike natural lan-
guage. The most widely used and understood logic is First-Order Logic (FOL) [146], also
known as First-Order Predicate Logic (FOPL). Whereas propositional logic assumes the
20
Contents
Chapter 2
First-order logic facts, objects, relations true / false / (unknown)
Temporal logic facts, objects, relations, times true / false / (unknown)
Probability theory facts degree of belief ∈ [0, 1]
Fuzzy logic facts with degree of truth ∈ [0, 1] known interval value
Table 2.1: Formal Logics and Their Ontological and Epistemological Commitments [146].
world contains facts therefore supports only one abstract relation ‘influences’, first-order
logic (like natural language) assumes the world contains (1) Objects, e.g. people, houses,
numbers, colors, baseball games, wars, etc. (2) Relations, e.g. red, round, prime, brother
of, bigger than, part of, between, etc. (3) Functions, e.g, father of, best friend, one more
than, plus, etc. For example, "Brother (Richard, John) ∧ Brother (John, Richard)" means
that "Richard is the brother of John and John is the brother of Richard"; ∀x King(x) =⇒
Person(x)" means that "All kings are persons". Logics of various kinds and logical reason-
ing and representation languages such as Prolog and KL-ONE have been popular tools for
knowledge modeling, for example in the definition of ‘expert systems’. The ontological and
A semantic network [141, 84] is a directed graph consisting of vertices that represent con-
cepts and edges that encode semantic relations between them. Concepts can be arranged
into taxonomic hierarchies and have associated properties (e.g. the state "Bill is a person"
could be represented by the chain: Bill Clinton Node - Is-A Arc - Person Node). In spite of
their simplicity and support for modular inheritance, semantic networks suffer from limited
expressiveness, as they can not represent negation or disjunction, among others. Frames
[118, 84] are closely related to semantic networks but represent knowledge in terms of hi-
erarchies of frames containing named slots, together with rules such as type constraints,
to define concepts and relationships. It is widely accepted that knowledge in the form of
21
Contents
semantic networks, frames, and scripts can be expressed using logics such as first-order
logic. Much recent work on knowledge modeling has focused on description logics [24, 2],
which have evolved from semantic networks to provide a more rigorous definition of seman-
tics in terms of a (typically restricted) form of first-order logic. Description logics usually
provide a syntax that makes it easy to specify categories and perform inference tasks such
The term ontology has recently undergone a strong revival largely due to the efforts of
the semantic web community [7]. Ontologies are seen in a knowledge management sense
as providing an important tool for the representation of semantic information and the
automated processing of such information for applications such as data mining, retrieval,
and automated discovery and utilization of services by autonomous software agents. The
shared conceptualization’ [75], hence the focus is on ‘knowledge sharing’ and ‘sufficiently
As shown in Figure 2.2, the semantic web is based on a layered hierarchy in which
[158, 14]. In addition to the XML and RDF standards for structured document annotation
and resource description, attention is now focussed on the new framework for Web On-
tology languages (OWL) [76]. It is now recognized that ontologies are the natural vehicle
for knowledge representation and interchange at different levels of granularity and across
different domains, hence they are to form a vital cornerstone of future generations of the
internet.
While the proposed formalisms for the semantic web draw on a rich heritage of work
in artificial intelligence and linguistics, they remain limited due to an explicit focus on the
into knowledge engineering have shown the limitations of methodologies such as semantic
22
Contents
Chapter 2
Figure 2.2: Semantic Web Architecture Layer Cake [14].
nets and description logics on which OWL is based. The World Wide Web Consortium
(W3C) and the knowledge engineering community have yet to address the enormous chal-
lenge of reasoning about non-textual multimedia data and supporting uncertain knowledge
representation. Some efforts have been paid to extend ontology language with uncertainty
representation capability using for instance fuzzy logic [38, 160, 181]. However, there is
no recommended languages for this purpose. So is remained at rule description and logic
framework layer. Even though there is several language candidates for describing rules
such as SWRL and RIF, they are still a working draft in the W3C and not yet a recom-
mendation. Logic framework layer that is assumed to provides more epistemic answer for
the question of why this piece of information is taken or appear to user ? again there is no
technology specification at present for this layer. Proof layer is assumed to answer agents
about the question of why they should believe the results. At present, again there is no
23
Contents
Overall, concerning our purpose of this dissertation, however the semantic web lay-
ered architecture gives us an insight on the component for epistemic reasoning framework.
Therefore, we regards the layered architecture rather as thorough requirements for intel-
ligent systems. At the level of metadata representation, there have been some efforts on
building ontological annotation taxonomies for multimedia description and event descrip-
tion. VERL (Video Event Representation Language) [122, 32] provides taxonomies to
annotate instances of the events and aceMedia project [28] aimed at providing semantic
taxonomies for multimedia annotation. In spite of such efforts, it is not yet widely ac-
cepted for annotation and reasoning due to the lack of handling uncertainties and lack
of technical support on logical reasoning and lack of technical means for proof layer as
explained above. Indeed, the focus of such efforts are to share semantic annotation in an
interoperable manner among systems, rather than focusing on the reasoning power.
Given a knowledge representation, reasoning is the process to draw new knowledge out of
observed patterns of facts and their attributes. Such reasoning methods can be categorized
into two approaches ‘extensional’ and ‘intensional’. These terms can be explained in
terms of defining sets. It is well known that there are two fundamental ways of defining
sets. First, a finite set may be defined by simply listing its members, as in the concept
may be defined by stating the properties of its elements. This is definition by ‘intension’.
with enumerating the final result and not other aspects of computation such as the path
traversed during computation, the amount of space or time required, etc [161]. In formal
logic and its realization such as ‘rule-based systems’ (a.k.a production systems), its primary
aim is to find ‘extensions’ based on their internal ‘search’ mechanism upon input facts.
24
Contents
such internal details in the given knowledge model [161]. These two classifications are
relative more than absolute, as the definition of ‘final result’ and ‘internal details’ may
Chapter 2
vary depending on the purpose of the semantic models and reasoning aspect.
logic as a means of representing knowledge and reasoning ‘extensionally’ [130], much of the
recent work in machine learning [62, 130] can be seen as providing ‘intensional’ definitions
that category and do not incorporate a description of the properties of the object. In fact
many statistical learning paradigms such as support vector machines rely on simplifying
and abstracting the classification problem to minimize the amount of information about
the problem that needs to be represented while maximizing the separability of the different
classes. A purely ‘extensional’ representation on the other hand consists of rules or state-
ments whose truth may be assessed independently of the particular extensions involved,
In the following we will briefly review ‘intensional approaches’ and ‘extensional ap-
proaches’ especially in consideration how these approaches handle uncertainty and seman-
tics.
to ‘states of affairs’ or subsets of ‘possible worlds’. Such systems treat uncertainty by con-
nectives that combine sets of worlds by set theory operations. For example, the probability
P (A ∧ B) is given by the weight assigned to the intersection of two sets of worlds, those
in which A is true and those in which B is true, but P (A ∧ B) can not be determined
from the individual probabilities P (A) and P (B). In intensional systems, the rules denote
m
elastic constraints about the world. For example, the rule A −→ B does not describe how
25
Contents
an agent reacts to the finding of A, but asserts that the set of worlds in which A and
¬B hold simultaneously has low likelihood and hence should be excluded with probability
m
m. In the Bayesian formalism the rule A −→ B is interpreted as a conditional probability
expression P (B|A) = m, stating that among all worlds satisfying A, those that also satisfy
certainty as a generalized truth value; that is, the certainty of a formula is defined to be a
unique function of the certainties of its sub formulas. Thus, the connectives in the formula
used in MYCIN [36] is a well-known example of ‘extensional system’. For example, the
assigned to A and B individually. The rules in extensional systems provide license for
m
certain symbolic activities. For example, a rule A −→ B may mean “If you see A, then you
are given the license to update the certainty of B by a certain amount which is a function
of the rule strength m”. The rules are interpreted as a summary of past performance of
the problem solver, describing the way an agent normally reacts to problem situations or
to items of evidence.
Consider Figure 2.3 that depicts the combination functions that apply to serial and
parallel rules, from which one can form a ‘rule network’. In extensional approach, it
given the credibility of each rule and the certainty of the premises (i.e., the roots of the
rule network). To complete the calculus we also need to define combining functions for
26
Contents
Chapter 2
Figure 2.3: An Example of Handling Uncertainty in Extensional Approach - Certainty
Combination Function used in MYCIN [130].
that the same combination function applies uniformly to any two rules in such systems,
‘regardless of what other rules might be in the neighborhood’. This is mainly due to the
nature of ‘modularity’ of inference in classical logic. For example, the logical rule “If A
• locality : If you see A anywhere in the knowledge base, then ‘regardless of what other
things’ the knowledge base contains, you are given the license to assert B and add it to
the database.
• detachment : If you see A anywhere in the knowledge base, then regardless of how
A was derived, you are given the license to assert B and add it to the database.
27
Contents
When it is considered with uncertainty, the procedural license provided by the rule
x
A −
→ B reads as follows : “If you see the certainty of A undergoing a change δA , then
regardless of what other things the knowledge base contains and regardless of how δA was
triggered, you are given an unqualified license to modify the current certainty of B by some
not give us license to do anything. Even if we are fortunate enough to find A true in the
database, we still can not assert a thing about B or P (B), because the meaning of the
statement is “If A is true and A is the ‘only’ thing that you know, then you are given license
to attach a probability p to B.”. As soon as other facts K appear in the database, the license
The probability statement leaves us totally impotent, unable to initiate any computation,
unless we can verify that everything else in the knowledge base is irrelevant. This is why
when relevance considerations are ignored. Such semantic deficiencies are posed :
‘detachment’ can create problems. In deductive logic the following holds true: A → B
by chaining. Derived evidence B triggers the rule B → C with the same rigor as would
a directly observed proposition. However consider the case, “ground is wet → it rained”
and ”sprinkler is on → ground is wet”. In this case, if an extensional system is told that
sprinkler is on, it will conclude that it rained. This is incorrect and in fact finding that
the sprinkler was on should only reduce the likelihood that it rained.
then finding B to be true makes A more credible (abductive reasoning). This requires
28
Contents
reasoning both ways. Extensional systems do not allow such bidirectional inference i.e.
Chapter 2
one has to explicitly specify the reverse rule, possibly risking creation of a cycle that
can cause evidence to be cyclically amplified until both cause and effect are completely
certain with no apparent factual justification. Removing the predictive component pre-
vents system from exhibiting another important pattern of plausible reasoning called
less credible. To exhibit this kind of reasoning, the system must use bi-directed infer-
ences; from evidence to hypothesis and from hypothesis to evidence. While it might be
possible to get around this problem by exhaustively listing all possible exceptions, to re-
store explaining away (without the danger of circular reasoning), any system that does
that sacrifices on principles of modularity. Inversely, any system that updates beliefs
modularly and treats all rules equally is bound to defy patterns of plausible reasoning.
result, they risk treating correlated evidence as independent. Consider a situation where
someone hears a piece of news independently from the radio, television as well as the
newspapers. Since from his point of view, the sources are independent, his belief in the
veracity of the piece of news should be very high. However, if that person were to realize
later that all the three sources got their information from the same source, then his belief
in the piece of news should decrease. This can never happen in extensional systems as
remedies for the problem of ‘bidirectional reasoning’ and ‘limitation of modularity’. The
way and the latter was addressed in the sense of ‘meta-reasoning’ using ‘abductive logic
programming’ [50].
29
Contents
Video
This chapter examines some relevant research trends and presents an overview of published
work that is related to automatic interpretation of high level semantic occurrence in video
sequences. We have chosen to organize semantic models into three different categories :
‘pattern recognition methods’, ‘state models’ and ‘semantic models’. ‘pattern recognition
methods’ do not generally address the representation aspect of semantic models and rather
trained and specified by feature vectors of training data. ‘state models’ represent domain
knowledge in a graphical model. ‘semantic models’ uses formal languages for knowledge
representation. The ‘state models’ and ‘semantic models’ categorizations again fall into
will further take a more in-depth look at the three categories of semantic interpretation
This class of techniques in this section is not quite semantic models, in the sense that they
do not consider the problem of semantic representation. Instead, they focus on the event
vector machines, neural networks, nearest neighbor classifiers, etc., are applied to the per-
ception scheme. Minimal semantic knowledge is needed in building the semantic classifiers
in this category.
The main advantage of the classifiers in this category is that they are well under-
stood. Usually, they may be fully specified from a set of training data. These approaches
are usually simple and straightforward to implement. This simplicity is afforded by ex-
cluding semantics (i.e., high-level knowledge about the semantic domain) entirely from the
30
Contents
There are many examples of pattern-recognition methods especially for event recog-
nition in the literature including [29, 180, 27, 71] and [151] (nearest neighbor), [136, 63, 135]
Chapter 2
(Support Vector Machine), [124, 157, 145] (boosting) and [167] (neural networks). A more
comprehensive discussion of these approaches and examples can be found in [12, 102, 165,
86].
“State” based semantic models are a class of formalisms that model the state of the video
semantics in space and time using semantic knowledge. State models improve on pattern-
recognition methods in that they ‘intrinsically’ model the structure of the state space of
the semantic domain. Such approaches fall into the category of ‘intensional’. Modeling
formalisms in this category are also well studied and mathematically well formulated.
This allows for efficient algorithms and sound formulations of problems such as parameter
learning. In most cases, however, the semantic information associated with the model
structure makes this structure difficult to learn from training data. In such approaches,
states are considered as symbolic facts and contextual knowledge is represented as a graph
structure having state nodes that are connected to each other. In the sense of logic,
connected two state nodes can be interpreted as a propositional logic rule that can consider
only one relation, ‘the causality implication’ (that can be often interpreted as ‘influence’,
exist within the whole graph structure, thereby, once uncertainty propagation mechanism is
learnt, adding additional pieces of knowledge will require restructuring causality influence
relation of the whole graph structure. This aspect restricts expressive power and increases
the modeling cost. Due to this complexity and lack of modularity, such approaches have
been focusing on relatively narrow and specific semantics. State modeling formalisms also
include FSMs (Finite-State Machines), Bayesian networks (BNs), hidden Markov models
FSMs assume a ‘fully observable sequence of states’, therefore have been used for
31
Contents
modeling the temporal aspects of video semantics with less concerns on uncertainty. FSMs
appearing in the literature naturally model single thread ‘event’ formed by a sequence of
states. FSM semantic models are utilized include hand gestures [91], single-actor actions
In order to deal with the inherent uncertainty of observations and interpretation that
exists in video, semantic models utilizing probability have been proposed. One such event
modeling formalism is the BNs (also known as probability network) [90]. In BNs, states
are considered as a random variable and they are connected with acyclic edges. Edges
represent joint probability between states using the notion of conditional independence.
BN semantic models are utilized for indoor surveillance [109] such as left luggage detection,
Hongeng et al. [85] considers an activity to be composed of action threads and recognizes
Remagnino et al. [144] uses BNs for parking lot surveillance, Wang et al. [170] uses a
BN model for primitive activity detection such as ‘jay walking’. One major drawback of
BN based semantic models is that they do not have an inherent capacity for modeling
problem include single-frame event classification [37] and choosing abstraction schemes
The benefits of a temporal evolution model (like FSM) and a probabilistic model
(like BN) are combined within the framework of the hidden Markov model (HMM). HMMs
are a class of directed graphical models extended to model the temporal evolution of the
state [142]. Due to this aspect, HMMs have become one of the most popular formalisms for
modeling video semantics. Makris et al. [113] uses HMM to reason about human behaviors
based on trajectory information. Oagale et al. [127] models single-person activities such
as ‘walking’ and ‘kneeling’. Gong et al. [72] applies a HMM model for airport tarmac
surveillance. Oliver et al. [128] use a layered HMM (LHMM) model in the event domain of
office surveillance. Due to the markov assumption that the current state depends only on
the state at a previous time, the semantics recognized in these works are tend to be mostly
32
Contents
a few seconds in length. One drawback of HMM is that as the model topology becomes
more complex, the efficient exact algorithms associated with the HMM structure are no
Chapter 2
longer applicable and must be replaced with approximation algorithms.
Dynamic BNs (DBNs) generalize BNs with a temporal extent. In fact, HMMs are a
special case of DBNs in which the structure is restricted to provide efficient algorithms for
learning and inference. This, however, often comes at the cost of computational tractability.
Approximation techniques are usually used to perform learning and inference. Thus, DBNs
in their general form appear less often as semantic modeling formalism in the literature.
[120] and [121] apply DBNs for surveillance of people such as ‘entering’, ‘passing’, etc.
resentation ‘scheme’ for surveillance systems [34], AND/OR tree for the analysis of specific
situations [46], or a GMM based scene representation for reasoning upon activities [114].
These extensions to the formalism have attempted to introduce aspects such as hier-
archy and uncertainty. These methods have largely been applied to specific event domains
While semantics such as many types of events can be described as a sequence of a number
of states, an interesting subset of semantics are those defined by the semantic relationships
level semantic models” groups several primitive unit semantics to allow these kinds of re-
lationships to be represented and recognized. These approaches do not aim to define the
‘entire state space’ of desired semantic domain as in ‘state modeling’ (intensional) ap-
proaches. Instead, semantic knowledge is used to define a set of semantic rules, constraints
and relations.
This type of approach allows the event model to capture high-level semantics such
as long-term temporal dependence (e.g. meeting, eating, etc.), hierarchical semantics (e.g.
boss of, employee of, etc.), semantically complex relations (e.g. friends of, thief of) using
33
Contents
results). Thus, approaches in this category are usually applied in domains where the
semantics (e.g. events, human activities or human interactions) of interest are relatively
complex and a particular semantics has large variance. However, because of the high-
level nature of this class of models, they often must be manually specified by a domain
expert (i.e., learning model structure and parameters is generally infeasible). Generally,
the formalisms in this category of event models such as grammars and logic formalisms
To cope with uncertainty, there have been some work on the use of logic program-
ming languages to achieve better expressive power and on the use of different uncertainty
handling formalisms to reason under uncertainty. Such logic framework based uncertainty
Yuri et al. [88] use a stochastic grammar and its parser for parking lot surveillance.
Ogale et al. [126] also use a stochastic grammar for human activity recognition. However,
in logical sense, their grammar rules corresponds to propositional logic, therefore has much
overlap with graphical ‘intensional’ models in the sense that it can not really represent
predicates as first-order logic. In fact, the achievement of better expressive power in ‘exten-
sional’ approaches is mainly due to the first-order predicate logic that logic programming
provides. While propositional logic deals with simple declarative propositions, first-order
logic additionally covers predicates and quantifiers. Akdemir et al. [13] proposed an ontol-
ogy based approach for activity recognition, but without uncertainty handling mechanism
(In ontology community, Description Logics (DLs) are often used as knowledge represen-
tation formalism and DLs are decidable fragments of first-oder-logic [24, 2]). Shet et al.
[152] proposed a system that adopts Prolog based logic programming for high-level reason-
ing. In [155] the same authors extended their system with the bilattice framework [69] to
perform the task of detecting humans under partial occlusion based on the output of parts
based detectors. Jianbing et al. [110] used rule-based reasoning with Dempster Shafer’s
Theory [149] for a bus surveillance scenario. Anderson et al. [17] used Fuzzy Logic [177] to
34
Contents
Chapter 2
Knowledge Rule Based Ontology Rule Based Rule Based Rule Based Rule Based
Modeling (Prop. L) (DL) (FOL) (FOL) (Prop. L) (FOL)
Traditional √ √ √
- - -
Logic Operators
Arithmetic √
- - - - -
Operators
Info. Fusion √ √ √
- - -
Operators
Extra Operators √
- - - - -
MT, MP,Rep,etc.
√
Default √
- - - -
Reasoning Chapter 5
√
Vague Rule √
- - - -
Modeling Chapter 6
√
Bidirectional
- - - - -
Inference Chapter 7
√
Diagnostic
- - - - -
Abduction Chapter 8
√
Belief √
- - - -
Revision Chapter 5-8
model human activity for video based eldercare. Dorado et al. [54] also applied fuzzy rules
but directly to low-level features for annotation of broadcasting contents. Therefore, their
focus was more on ‘vision based perception’ aspect. They defined propositional logic based
of low-level features and ‘<action>’ denotes annotating a word (concept). For example,
using the MPEG-7 edge histogram descriptor (EHD), they mapped an EHD to a fuzzy set
BU ILDIN G’. Such V , H, D and N D can be considered as state symbols that can be
35
Contents
posed by low-level features. Because their rules are propositional it can be also represented
as a graph having such symbols as nodes that are connected to a word symbol. In this
In this dissertation we proposed the use of logic programming and subjective logic
[93] to encode contextual knowledge with uncertainty handling. Based on our previous
work [78, 81, 80, 82, 83, 79], we extend the system to demonstrate bidirectional condi-
tional inference, belief revision, default reasoning and vague rule modeling. Table 2.2
shows a brief comparison of the previously proposed ‘extensional’ approaches. The ta-
ble shows that the coverage of our subjective logic based approach is most broad. For
example, while some provides information fusion capability for fusing two contradictory
information sources, such as dempster shafer’s fusion operator, bilattice’s operator and
subjective logic’s consensus operator, only some of them support default reasoning that
handles such contradictory information to draw reasonable decision and belief revision.
Indeed, our system also supports modeling vague propositional rules and inference under
such vague rules for high level semantic analysis of visual surveillance data (in the sense
of linguistic interpretation of the rules, the most similar previous approach to the pro-
posed work would be [17]). Finally, bidirectional inference (as is possible in ‘intensional’
36
3 Preliminaries
Chapter 3
This chapter gives an overview of the fundamental background about subjective logic
theory and logic programming that will be discussed throughout this dissertation.
Audun Jøsang introduced subjective logic as a framework for artificial reasoning [92, 93].
Unlike traditional binary logic or probabilistic logic (the former can only consider true or
false, and the latter can consider degrees of truth or falseness), subjective logic explicitly
represents the amount of ‘lack of information (ignorance) on the degree of truth about a
proposition’ in a model called ‘subjective opinion’ [92]. The idea of explicit representation
of ignorance is inherited from the Dempster Shafer belief theory [149, 92, 149] and the in-
distributions [93]. Subjective logic also comes with a rich set of operators for the manipu-
lation of opinions. In addition to the standard logical operators, subjective logic provides
some operators specific for Dempster Shafer belief theory such as consensus and recom-
mendation. However, unlike Dempster Shafer’s evidence fusion rule that is inconsistent
with Bayes theorem, it provides an alternative consensus rule with a solid mathematical
basis [92]. It is also different from fuzzy logic: While fuzzy logic maps quantitative mea-
sure to non-crisp premises called fuzzy sets (e.g. ‘fast’, ‘slow’, ‘cold’, ‘hot’, etc.), subjective
logic deals with the uncertain belief itself on a crisp premise (e.g. ‘intrusion happened’,
into the linguistic certainty fuzzy set (i.e., ‘very certainly true’, ‘less certainly true’, etc)
37
Contents
is also possible. In general, subjective logic is suitable for modeling real situations under
Known application areas are trust network modeling, decision supporting, modeling
and analyzing Bayesian network, etc. However, to the best of our knowledge, the applica-
tion of subjective logic in computer vision related domains has been limited to our previous
work [78, 81, 80, 82, 83, 79]. In this dissertation, based on our previous work, we demon-
interpretation of conditional rules using abduction and deduction operators and modeling
vague rules relying on reputation operator in subjective logic. In this section, we will give
Subjective logic uses theoritical elements from the Dempster-Shafer belief theory [149].
In this section, we give a brief introduction to Dempster Shafer belief theory with an
example. Dempster Shafer belief theory deals with a set of hypotheses called the ‘frame
mutually exclusive, therefore exactly one element is assumed to be true. To illustrate the
frame of discernment let us consider following example shown by Zadeh [178] (The same
case with three suspects; Peter, Paul and Mary then Figure 3.1 is an example of frame
of discernment shown in [94]. If an element is assumed to be true, its supersets are also
considered to be true as well. An observer can assign belief mass to one or several states
38
Contents
Chapter 3
that one of the elements in x4 is true but the observer is uncertain about which of them is
true (i.e., ignorance). In contrast to belief mass, the ‘belief ’ to a set x should be the sum
BMA on Θ. Then the belief function corresponding with mΘ is the function b : 2Θ → [0, 1],
∑
such that b(x) = y⊆x mΘ (y), where x, y ∈ 2Θ .
This idea can be expanded to the cases of ‘disbelief ’ and ‘ignorance’ as follows.
By the Definition 1, the sum of the ‘belief ’, ‘disbelief ’ and ‘ignorance’ is equal to the
sum of the belief masses which sums up to 1. Therefore, b(x) + d(x) + i(x) = 1. Given the
elements in frame of discernment, the relative number of atomic elements is called ‘relative
39
Contents
mA
Θ bA (x) dA (x) iA (x) mB
Θ Dempster’s Non-normalised
x∈2 Θ
rule Dempster’s rule
Peter 0.98 0.98 0.01 0.01 0.00 0.490 0.0098
Paul 0.01 0.01 0.98 0.01 0.01 0.015 0.0003
Mary 0.00 0.00 0.99 0.01 0.98 0.490 0.0098
Θ 0.01 1.00 0.00 0.00 0.01 0.005 0.0001
∅ 0.00 0.00 1.00 0.00 0.00 0.000 0.9800
lowing definition.
BMA mΘ , then the probability expectation function corresponding with mΘ is the function
∑
E : 2Θ → [0, 1] such that, E(x) = y mΘ (y)a(x/y), where y ∈ 2Θ .
In Dempster Shafer belief theory, belief mass assignment mΘ plays the most fun-
damental basis for representing belief and calculating probability expectation, etc. In
addition, Dempster Shafer belief theory comes with a rule to combine two different belief
′ B A ′ B
∑
function mAΘ ⊕ mΘ : 2 → [0, 1] such that, 1. mΘ ⊕ mΘ (∅) =
Θ A B
y∩z=∅ mΘ (y)·mΘ (z)−Kand
∑
′
A B
y∩z=x mΘ (y)·mΘ (z)
∑
2. mAΘ ⊕ mΘ (x) = , for all x ̸= ∅, where K = y∩z=∅ mA Θ(y) · mB
B
1−K Θ (z)
of Dempster’s rule.
Suppose that two observers assigned their belief masses to the frame of discernment
shown in Figure 3.1 that its elements are suspect of a murder case. Table 3.1 shows an
40
Contents
The ideas of Dempster Shafer belief theory can be simplified in subjective logic theory, by
restricting the frame of discernment to be binary, i.e., it will only contain (focus on) one
particular set and its complement. Such a frame of discernment is called ‘focused frame
of discernment’. Figure 3.2 shows examples of focused frame of discernment that can be
derived from Figure 3.1. The formal definition of ‘focused frame of discernment’ is as
Chapter 3
follows.
and let x ∈ 2Θ . The frame of discernment denoted by Θ̃x containing only x and ¬x(i.e., x,
focus on x.
Remember that the belief, disbelief and ignorance functions are also indexed by a
specific element in a frame of discernment (i.e., b(x), d(x), i(x), respectively). By this
ment with belief mass assignment mΘ where b(x), d(x) and i(x) are the belief, disbelief and
ignorance functions of x in 2Θ , and let a(x) be the real relative atomicity of x in Θ. Let
Θ̃x be the focused frame of discernment with focus on x. The corresponding focused belief
mass assignment mΘ̃x and relative atomicity aΘ̃x (x) on Θ̃x is defined according to:
m x (x) = b(x)
Θ̃ a x (x) = E(x)−b(x) , f ori(x) ̸= 0
Θ̃ i(x)
mΘ̃x (¬x) = d(x) .
a (x) = a(x), f ori(x) = 0
Θ̃x
mΘ̃x (Θ̃x ) = i(x)
represented based on the concept of focused belief mass assignment as shown in Figure 3.3.
41
Contents
Figure 3.2: Examples of Focused Frame of Discernment Derived from Figure 3.1.
Definition 10. (Opinion) [93] . Let Θ = {x, ¬x} be a state space (a.k.a ‘Frame’)
containing x and its complement x. Let bx , dx , ix represent the belief, disbelief and ig-
norance in the truth of x satisfying the equation: bx + dx + ix = 1 and let ax be the base
rate of x in Θ. Then the opinion of an agent ag about x, denoted by wxag , is the tuple
wxag = (bag ag ag ag
x , dx , ix , ax ).
In the context of the frame of discernment shown in Figure 3.1 and belief mass
assignment shown in Table 3.1, the examples of the focused opinions of agent A are wPAeter =
(0.98, 0.01, 0.01, 1/3), wPAaul = (0.01, 0.98, 0.01, 1/3) and wM
A
ary = (0.00, 0.99, 0.01, 1/3).
The probability expectation can be also defined on the opinion model in subjective logic.
opinion about the truth of x, then the probability expectation of wxag is defined by: E(wxag ) =
bag ag ag
x + ax ix .
A point inside the triangle represents a (bx , dx , ix ) triple. The corner points marked
with Belief, Disbelief and Ignorance represent the extreme cases, i.e., full belief (1, 0, 0),
full disbelief (0, 1, 0) and no knowledge (0, 0, 1). An opinion lying on base line (the line
connecting Belief and Disbelief) is called ‘dogmatic opinion’ in the sense that they do not
contain any degree of ignorance (i.e., ix = 0). Such dogmatic opinions correspond to
traditional traditional probability. The base rate ax represents the prior knowledge on the
tendency of a given proposition p’s being true and can be indicated along the base line.
42
Contents
Chapter 3
Figure 3.3: Opinion Triangle.
Figure 3.4: Linguistic Fuzzy Category for Opinion Triangle as a Function of the Base Rate
[139].
43
Contents
For example, in the case we toss a balanced coin with the proposition x as ‘get head’,
then we put ax = 1/2 however, in the case of biased coin we could set different values.
Similarly, in the case we toss a balanced dice we put ax = 1/6 (note that, the original
frame of discernment has six elements). Usually, when we consider balanced binomial
cases, the default value is 1/2. The probability expectation E is then formed by projecting
the opinion onto the base line, parallel to the base rate projector line (see the blue line)
that is built by connecting the ax point with the Ignorance corner (see the red line).
and uncertainty, a linguistic interpretation of an opinion is also possible. Pope et al. [139]
showed an example mapping of fuzzy categories to opinion triangle. As shown in Figure 3.4,
the stronger opinion we have, the lesser changes of mapping areas are made. Similarly, the
weaker opinion we have, we are relying the more on base rate. For example, the area 5A
can be interpreted linguistically as ‘completely certain chances about even’ and the area 2B
As shown in the Definition 10, subjective opinion’s frame Θ is binomial, because it deals
with two elements about a proposition x. Namely, x (i.e., x happened, success) and x (i.e.,
x does not happened, failure). Assuming n number of independent observations about the
proposition x with a fixed base rate ax = π, the conditional probabilistic likelihood that x
we are holding ax fixed and are looking at the probability distribution of y over its possible
discrete integer values of 0, ..., n. This is referred to as ‘binomial distribution’ and noted
number of successes y, holding y fixed and let π vary over its possible values 0 ≤ π ≤ 1.
44
Contents
Taking a different view, let us assume that we want to know about the probability density
of the probability π itself given y successes of observations, namely g(π|y). By the Bayes’
theorem, we know that ‘posterior ∝ prior × likelihood’, therefore, g(π|y) ∝ g(π) × f (y|π).
This gives us only the shape of the posterior density. To get the actual posterior, we
need to normalize this by some constant k, to make sure that the area under the posterior
integrates to 1. We find k by integrating g(π) × f (y|π) over the whole range. In general
Chapter 3
we get,
g(π) × f (y|π)
g(π|y) = ∫ 1 , (3.1)
0 g(π) × f (y|π) dπ
In above Equation 3.1, it requires an integration. However, depending on the prior g(π)
chosen, we do not always need to do the actual integration numerically. Assume that we
do not have any idea beforehand what the π is, therefore, we assign,
Clearly, we see that in this case, the posterior density is proportional to the likelihood
factorial part in the coefficient can be generalized to the cases of real values using the
Γ(α + β) α−1
g(π|α, β) = π (1 − π)β−1 = Beta(π|α, β), f or 0 ≤ π ≤ 1 (3.3)
Γ(α)Γ(β)
Above Equation 3.3 is the probability density function called beta distribution 2 that its
shape is only dependent on the indexes α and β. Now, let us consider assigning the beta
1
The Gamma function is a generalization of the factorial function over integer ∫to real and complex
∞
numbers, with its argument shifted down by 1. Formally, it is defined as Γ(z) = 0 tz−1 e−t dt.
2 1
Note that, the coefficient part of the beta distribution is B(α,β) , where B(α, β) is a special function
∫1
called Beta function that is formally defined as B(α, β) = 0 tα−1 (1 − t)β−1 dt = Γ(α)Γ(β)
Γ(α+β)
.
45
Contents
Γ(α + β) α−1
Using a Beta Prior : g(π; α, β) = π (1 − π)β−1 , f or 0 ≤ π ≤ 1. (3.4)
Γ(α)Γ(β)
the Bayes’ theorem. To make g(π|y) a probability density function that the area under
k × π y+(α−1) (1 − π)n−y+(β−1)
g(π|y) = ∫ 1 = k ′ × π y+(α−1) (1 − π)n−y+(β−1) . (3.5)
0k×π
y+(α−1) (1 − π)n−y+(β−1) dπ
In the above Equation 3.5, we can easily recognize that this can be a beta distribution
Γ(n + α + β)
g(π|y) = π y+α−1 (1 − π)n−y+b−1 , f or 0 ≤ π ≤ 1. (3.6)
Γ(y + α)Γ(n − y + β)
Again, the posterior density of π has been easily obtained without having to go through
the numerical integration. Furthermore, the posterior is to be the same form as the
power. When we multiply the beta prior times the binomial likelihood, we just add the
exponents of π and (1 − π), respectively. This makes using Beta(π|α, β) priors when we
have binomial observations particularly easy. Such priors that make their posteriors of the
same form, is called ‘conjugate family of priors’. Such ‘conjugate priors’ play an important
our discussion, we explained that beta distribution is the conjugate family for the ‘bino-
mial’ observation distribution. Another advantage of the beta distribution is that it offers
various shapes according to its indexed parameters α and β as shown in Figure 3.6. In fact,
the uniform distribution g(π) = 1 used in (3.2) is a special case of beta distribution that is
indexed with α = 1 and β = 1, namely, Beta(π|1, 1) (see Figure 3.6-(b)). To explain this
46
Contents
Chapter 3
Figure 3.5: Examples of Beta Distributions.
Figure 3.6: Examples of Applying Bayes’ Theorem to Different Priors and Given a Likeli-
hood.
The essence of Bayesian statistics, as explained above, is the use of both the prior
knowledge and the experimental result called likelihood, to get the posterior knowledge.
For example, assume that we have tossed a biased coin 5 times today and got 2 times of
head. Then, what is the chance that we would get head if we toss the coin tomorrow ?
To infer this, ideally, it would be good if we could appropriately fuse a subjective belief or
opinion about the coin and an objective experimental likelihood (2 times of head out of 5
47
Contents
trials). The subjective opinion in bayesian statistics is called ‘subjective probability’ [33]. In
general, the subjective probability is not a single probability value but a distribution over all
possible probability values. Therefore, the characteristics that beta family of distributions
can form a various shapes give us high degree of flexibility on modeling the desired prior
density. Further, it is also known that some of other types of well known priors can be also
approximated to beta distributions [33]. For instance, Alfers et al. introduced a normal
(gaussian) approximation for beta and gamma probabilities [9], Teerapabolarn introduced
a poisson approximation to the beta binomial distribution [164], etc. (refer to [33] for more
In bayesian statistics, however, the effect of the prior we choose will be small when
we have a likelihood from enough data. For example, consider three students Anna, Bart
and Chris are constructing their prior belief about the proportion of residents in their city
who support building a casino (this example is shown in [33]). Based on their subjective
beliefs, let us assume that they have chosen weak priors as shown in Figure 3.6 - (a). Then
assume that they took a random sample of n = 100. Out of the random sample, 26 said
they support and 74 said they do not support building a casino. Figure 3.6 - (b) shows
this likelihood. Despite starting with different priors, when they are considered with the
likelihood, the three posteriors shown in Figure 3.6 - (c) show very similar posteriors. In
other words, in general, putting roughly shaped ‘subjective prior’ is considered to be enough
in case we have reasonable amount of data or reasonably strong likelihood distribution [33].
Audun Jøsang introduced the mapping between beta distribution Beta(Θ|α, β) and
subjective opinion wx = (bx , dx , ix , ax ) [92, 98]. Consider a beta distribution over the frame
(state space) Θ and let π be the probability that x ∈ Θ could happen. Replacing r = α − 1
Γ(r + s + 2)
Beta(π|r, s) = π r (1 − π)s , f or 0 ≤ π ≤ 1, r ≥ 0, s ≥ 0. (3.7)
Γ(r + 1)Γ(s + 1)
48
Contents
Γ(r + s + 2)
Beta(π|r, s, ax ) = π r+2ax −1 (1 − π)s+2(1−ax )−1 . (3.8)
Γ(r + 2ax )Γ(s + 2(1 − ax ))
Chapter 3
(r+2ax )
Then the probability expectation of Equation 3.8 is known to be as E(π) = (r+s+2) .
Similarly, by the Definition 11, Ewx = bx + ax ix . Then we need to make the two notations
of E to be the same (i.e., E(wx ) = bx + ax ix = E(π) = (r + 2ax )/(r + s + 2) ).
bx +
a a
x ix = r/(r + s + 2) + 2x /(r + s + 2)
b +d +i =1
x x x
Now, remember that we have replaced the parameters α and β in Equation 3.7, with
α = r + 2 · ax and β = s + 2 · (1 − ax ), by this and above (3.9) we can get parameters r, s,
α and β as follows,
1. when ix ̸= 0,
r = 2bx /ix → α = 2bx /ix + 2 · ax s = 2dx /ix → β = 2dx /ix + 2 · (1 − ax )
1=b +d +i ax = base rate of x
x x x
49
Contents
Figure 3.7: Examples of Opinion Triangle and Their Mapping to Beta Distribution.
This means, for example, that an opinion with bx = 0, dx = 0, ix = 1 and ax = 0.5 which
maps to Beta(1, 1) is equivalent to a uniform probability density function (see Figure 3.6-
(b)). It also means that a dogmatic opinion with ix = 0 which maps to Beta(bx η, dx η)
where η → ∞ is equivalent to a spike probability density function (i.e., dirac delta) with
infinitesimal width and infinite height at bx (see Figure 3.6-(g) and (h)). Dogmatic opinions
in beta distribution perspective means that infinite amount of evidences converges to the
α 3
ratio of α+β [98].
can be interpreted as ‘seems likely and slightly uncertain true’, and example 2) shows full
ignorance (a.k.a. ‘vacous’ opinion) at the time of judgement about a proposition. Assuming
base rate to be 0.7 in the example we get expectation value also to be 0.7 and the beta
3
however, we introduce a bound to avoid the actual numerical computation of infinity value and to
visualize the distribution. Currently, we assign a very tiny value such as ix = 1 × 10−5 for numerical
calculation, when ix → 0.
50
Contents
distribution appears biased towards ‘True’ though the opinion represents full ignorance.
Subjective logic is a generalization of binary logic and probability calculus. This means
that when a corresponding operator exists in binary logic, and the input parameters are
equivalent to binary logic TRUE or FALSE, then the result opinion is equivalent to the
Chapter 3
result that the corresponding binary logic expression would have produced. Table 3.2 pro-
vides a brief overview of the main subjective logic operators. Additional operators exist for
modeling special situations, such as when fusing opinions of multiple observers. Most of the
operators correspond to well-known operators from binary logic and probability calculus,
whereas others are specific to subjective logic. For example addition is simply a generaliza-
tion of addition of probabilities. However, subjective logic also comes with non-traditional
logic operators. For example, deduction[95, 100], abduction[95, 139], discounting [99] op-
of different sources of observations. Cummulative (average) fusion (unfusion) [99, 95] oper-
ators are generalization that is inspired from the dempster shafer’s belief fusion rule [149]
Apart from the computations on the opinion values themselves, subjective logic op-
erators also affect the attributes, i.e. the subjects, the propositions, as well as the frames
containing the propositions. In general, the attributes of the derived opinion are functions
of the argument attributes. Following the principle illustrated in Figure 3.8. the de-
rived proposition is typically obtained using the propositional logic operator correspond-
ing to the subjective logic operator. For example, consider two frames of discernment
of X and Y will be the cartesian product of X and Y connected through one of above
logical operators, and fpl (x, y) will be one of possible elements in fF C . Similarly, fsc will
simply become A because in this case we are considering only one agent A. Therefore,
51
Contents
Table 3.2: Subjective Logic Operators, Notations, and Corresponding Propositional / Bi-
nary Logic Operators.
A
for instance, whealthy∧dayworker A
= whealthy · wdayworker
A deals with two focused propositions
‘healthy’ and ‘dayworker’ in X and Y . Then doctor A would have an opinion about the
proposition healthy ∧ dayworker to the amount of opinion that is calculated via subjective
The functions for deriving attributes depend on the operator. Some operators, such
as cumulative and averaging fusion, only affect the subject attribute, not the proposition
which then is equal to that of the arguments. Fusion for example assumes that two separate
argument subjects are fused into one. Other operators, such as multiplication, only affect
the proposition and its frame, not the subject which then is equal to that of the arguments.
Multiplication for example assumes that the derived proposition is the conjunction of the
argument propositions, and that the derived frame is composed as the Cartesian product
52
Contents
of the two argument frames. The transitivity operator is the only operator where both the
subject and the proposition attributes are affected, more specifically by making the derived
subject equal to the subject of the first argument opinion, and the derived proposition and
frame equal to the proposition and frame of the second argument opinion.
logic expressions as attributes of derived opinions. Instead, the trust origin subject and a
Chapter 3
compact substitute propositional logic term can be used.
form, or as [A, B] : [B, C] in expanded form. The expanded form is the most general, and
corresponds directly with the way subjective logic expressions are formed with operators.
and uncertainty handling are based on the extension of logic programming with subjective
logic formalism. While subjective logic also is a logic, it’s expressive power is remain within
dissertation, to cope with complex semantics, we adopt ‘predicate’ logic that is based on
A propositional logic is the most basic branch of mathematical logic in which the only
objects are propositions, that is, objects which themselves have truth values. Variables
for the constants T and F (representing true and false respectively). Propositional logic
53
Contents
disjunction and implication). A model for propositional logic is just a ‘truth function’ on
a set of variables. Such a truth function can be easily extended to a truth function on all
formulas which contain only the variables are defined on by adding recursive clauses for the
determine whether a sentence is a tautology (i.e., a proposition that is true for all possible
interpretations). It can be done using truth tables, since a truth table for a particular
formula can be easily produced, and the formula is a tautology if every assignment of
from a countable alphabet as defined in Definition 12, and formed according to certain
Definition 12. (The Alphabet for Propositional Formulae) . this alphabet consists
of:
(2) The logical connectives : ∧ (and), ∨ (or), → (implication), ¬ (not) and sometimes ≡
(equivalence).
(3) Auxiliary symbols : ‘(’, ‘)’ (left and right parenthesis respectively).
mulae (or propositions) is the inductive closure of the set P S ∪ F under the functions
C¬ ,C∧ ,C∨ ,C→ and C≡ , defined as follows : For any two strings A, B over the alphabet of
Definition 12,
C≡ (A, B) = (A ≡ B).
The truth table in propositional logic can be generalized into continuous space by
replacing the truth values and logical connectives with for instance, probability values
and probabilistic calculation. For instance, given propositional sentences A and B, · for
54
Contents
∧, + for ∨, complement for ¬ and P (B|A) for A → B. Similarly, subjective logic also
corresponds to propositional logic sense in that it deals with propositions. In this sense,
‘intensional’ approaches that models propositional logic sentences in a graphical model can
values.
Chapter 3
3.2.2 First Order Predicate Logic
The weak expressive power of propositional logic accounts for its relative mathematical
simplicity, but it is a very severe limitation, and it is desirable to have more expressive
logics. First-order predicate logic is a considerably richer logic than propositional logic, but
yet enjoys many convinient mathematical properties. For example, in propositional logic
the proposition ‘John is tall’ can not be decomposed into a simple sentence because there is
to decompose the proposition into predicates and individuals as ‘tall (john)’. First-order
expressions. For example, propositions ‘Every cat is sleeping’, ‘Some girls likes David’ or
nically, this is achieved by allowing the propositional symbols to have arguments ranging
over elements of structures. For convenience, we also allow symbols denoting functions and
constants.
Definition 14. (The Alphabet for First-Order Predicate Formulae) . this alphabet
55
Contents
(3) Auxiliary symbols : ‘(’, ‘)’ (left and right parenthesis respectively).
(i) Function symbols: A (countable, possibly empty) set F S of symbols f0 , f1 , ..., and
a rank function r assigning a positive integer r(f ) (rank or arity) to every function
symbol f .
(ii) Constants: A (countable, possibly empty) set CS of symbols c0 , c1 , ..., each of rank
zero.
(iii) Predicate symbols: A (countable, possibly empty) set P S of symbols P0 , P1 , ..., and
a rank function r assigning a nonnegative integer r(P ) (called rank or arity) to each
predicate symbol P .
over the alphabet A of Definition 14 is the collection of all WFFs (Well formed formulas)
Based on the formal definition given above, the aforementioned example propositions
Logic Programming is, in its broadest sense, the use of declarative mathematical logic for
more commonly understood, is the use of ‘predicate’ logic as both a declarative and proce-
dural representation language (in the sense in which it also supports procedural ‘predicate
functions’). Logic Programming considers logic theories of a specific form. The theories,
called logic programs, mainly consists of two types of logical formulae, rules and facts.
56
Contents
Rules are of the form ‘A ← f0 , f1 , ..., fm ’ where A is rule head and the right hand side is
called body. Each fi is an atom and ‘,’ represents logical conjunction. Each atom is of
the form ‘p(t1 , t2 , ..., tn )’, where ti is a term and p is a predicate symbol of arity n. Terms
gramming language, it also supports defining procedural ‘predicates functions’ and some
of reserved predicates are used as ‘predicate functions’ to check equality of the terms (e.g.,
Chapter 3
eq(t1 , t2 ) returns true when t1 and t2 are the same in CLIPS [1]) or to compare/calculate
arithmetic values between terms of predicates (e.g., +(t1 , t2 ) the reserved predicate ‘+’
acts as a function returns value of t1 + t2 , similarly, the functional predicate leq, leq(t1 , t2 )
returns true in case that the condition t1 ≤ t2 is satisfied in CLIPS). There is one special
term, called the anonymous term, for which the underscore (_) character is used. When
the character is used, it basically means that it does not care which variable or symbol it
just ‘f ’) is called facts and can serve as an atom when used in a rule body. Negation is
represented with the symbol ‘¬’ such that ‘¬¬A=A’. Both positive and negative atoms
are referenced to as literals. Given a rule ‘head ← body’, we interpret the meaning as ‘IF
body THEN head’. Traditionally, a resolved set of facts that matches to a rule is called
ones mentioned in Section 2.4.3, rules have been used to define and reason about various
of T . In other words, adding information never invalidates any conclusions. However, rea-
soning under ‘uncertain knowledge’ (as of our interest for reasoning in visual surveillance)
is different. We often draw plausible conclusions based on the assumption that the world
in which we function and about which we reason is normal and as expected. This is far
57
Contents
from being irrational. To the contrary, it is the best we can do in situations in which we
have only incomplete information. However, as unexpected as it may be, it can happen
that our normality assumptions turn out to be wrong. New information can show that
the situation actually is abnormal in some respect. In this case we may have to revise our
called ‘nonmonotonic’. It has been a focus of extensive studies by the knowledge represen-
tation community since the early eighties of the last century. This interest was fueled by
several fundamental challenges facing knowledge representation such as modeling and rea-
soning about rules with exceptions or defaults. Another important class of ‘nonmonotonic’
Therefore, according to the given set of observations such reasoning need to be able to rate
logic for uncertainty representation formalism. However, subjective logic itself is remain
this dissertation to cope with complex semantics, we adopt ‘predicate’ logic that is based
on first-order logic. For the flexible manipulation of facts and predicates on the need of
arithmetic calculation, etc., we also benefit from procedural handling of them. Therefore,
58
4 Proposed Architecture and Case Study on
Forensic Queries
Thus far, we have reviewed fundamental background of this dissertation. The discussions
covered paradigm changes in intelligent visual surveillance, technical challenges and re-
Chapter 4
quirements toward intelligent visual surveillance system, prior art and related work appear
in literature and important preliminaries. Bearing the background of this work in mind,
this chapter starts with the architectural aspect of our approach to high level semantic
4.1 Introduction
In this chapter, we present our conceptual system architecture supporting high-level seman-
tic analysis of visual surveillance data. Taking pragmatic view, efficient semantic analysis
of visual surveillance data requires a system oriented approach that optimally combines
the individual legacy vision analytic power and its optimal use. Regarding this objective,
5) Epistemic meta reasoning and abduction. Therefore, reminding the discussion, the
fall into two parts: ‘1) knowledge representation and reuse’ and ‘2) epistemic reasoning
mechanism’. In this chapter, we give more focus on ‘knowledge representation and reuse’
59
Contents
introducing a ‘data model’ for contextual metadata description. Regarding the ‘epistemic
logic’, and the detailed aspect on particular reasoning mechanisms will be focussed in the
rest of chapters.
We first start with discussion on the software components, data processing pipeline
and system architecture design. Then we introduce our prototype implementation and
present some of case study on forensic query scenarios. We conclude with discussions
Figure 4.1 shows the overall architecture proposed base on the discussed critical require-
‘heterogeneous vision analytics’. To resolve this issue, individual vision analytics should
share a principled description method. As introduced in Section 2.2.3, there have been
remarkable research efforts on the formal description for the sharable knowledge in the
field of ‘ontological knowledge description’ or ‘semantic web’. Following this recent trend,
representation: it provides the basis to enable fetching and collecting meaningful subsets
of information from unstructured data sets by giving specific and sharable semantics to
terminologies. Description Logic [24] and its encoding format RDF/OWL-DL [76] is a
good candidate due the widely available supporting tools (such as OWL-DL reasoners,
triple stores and SPARQL query APIs). Along the guideline of metadata ontology, in-
termediate video analysis results will get packaged as ontological metadata instances.
60
Contents
Chapter 4
Figure 4.1: Conceptual Architecture of the Proposed Intelligent Forensic Reasoning Sys-
tem.
Packaged metadata instances will get fed into the archival layer, where the low-level fea-
ture vectors are indexed in an appropriate index structure [31] and ontological instances
will be archived into triple store [103] backed with a relational data base (RDBMS).
difficulties to apply predefined state model based approaches, the logic programming
approach [104] is exploited (see Section 2.3 and Section 2.4 for the detailed comparison
the detailed procedures and instructions are omitted but what the machine should do
is described. The steps to solve the given problem are executed by a kind of a runtime
system (called rule engine or logical reasoner) that understands how to use the declarative
information. Therefore, the models for solving the problems are shifted from a set
a visual surveillance scenario can be represented as sets of rules. Each rule can be
61
Contents
without hitting whole knowledge structure as of state model based approaches (see the
Section 1.2.4, vision analytic metadata usually comes with ‘uncertainty’ and therefore
influence the acceptable accuracy and robustness for a semantically complex queries.
state-based models (such as FSMs, BNs, HMMs, DBNs, etc. see Section 2.4.2 for de-
tails.), and also considering the popularity of ‘state-based models’ in vision analytic
modules, the resulting values tend to represent probability of being true about target
semantics of vision analytic modules. Therefore, the values normally lay in the interval
[0, 1]. In logic programming, such metadata can be seen as facts (see Section 3.2.3) and
the facts should be fed into rule-engine with the values attached. Our approach is to
encode such values as subjective logic’s opinion [93]. Subjective logic is compatible to
the traditional probability theory but can also handle ‘uncertainty about the probability
can be seen as either true or false. As discussed above, however, the facts will come with
‘subjective opinion’ in our design. In addition, semantic knowledge encoded as rules can
not be considered as absolute ones. Rather such rules are also epistemically uncertain.
Thus, we need a means to represent the strength or belief on a particular rule as well.
As ‘opinion assignment to metadata’, we will also attach an epistemic belief to each rule
using subjective logic opinion. This way of ‘epistemic uncertainty representation’ will
(i.e., metadata in terms of vision analytics, and that can be seen as a fact in terms of
logic programming) and rules, an additional step is needed to deal with the uncertain
62
Contents
aspect of the gathered information and rules. For example, evaluating the amount of
information sources towards the same conclusion, the amount of information sources
it should be also possible to revise current belief upon arrival of new information and
aforementioned evaluations. We use subjective logic due to its rich set of logical oper-
ators such as consensus, discounting, averaging fusion etc., which are not available in
more details).
Chapter 4
• Epistemic Meta Reasoning and Abduction: As shown in Section 1.2.3, for large
reasoning) power that reason about the possible sub queries from a set of observation.
Thus, such reasoning should be done by setting possible semantic hypotheses upon given
observations and assessing each hypothesis in an iterative manner. This aspect resem-
bles diagnostic reasoning which the so called ‘abductive (constraint) logic programming’
[50] plays an important role. Therefore, we embed the capability of ‘abductive logic pro-
gramming’ inside our rule engine. Both normal logic programming based reasoning and
abductive logic programming lay in the whole reasoning pipeline to interact each other
in Section 1.2.4, one fundamental aspect of such a high level semantic analysis sys-
tem should deal with is the method to deliver user intention into a system. Machine-
format with appropriate query interface. Once a query is delivered, an agent should
decide where to access to fetch proper information in case there are several information
sources. We name this particular conceptual agent as ‘Query Broker’ (and a cooperative
project is undergoing on this issue). However, we will leave this particular component
63
Contents
out of scope in this dissertation but rather focus on above mentioned items.
As proof of concept, a demonstrator along the afore introduced architectural approach has
been built. In this section, we mainly focussed on metadata representation and inference
mechanism upon acquired metadata instance among all requirements. We first describe
the conceptual ontological metadata description model, Next the subjective logic extension
An ideal high-level semantic analysis of visual surveillance system should enable optimal
reuse of currently available assets of vision analytic power. In reality, various vision analyt-
ics could exist and deployed from varying vendors or legacy systems. In this sense, proper
reuse of intermediate results generated from each of individual vision analytic modules in-
issue, individual vision analytics should share a principled formal metadata description
method. This metadata description should be also considered in the sense of logical rea-
soning. The logical reasoning framework usually gets list of facts and then returns derived
knowledge. Moreover, each fact should be described with specific semantics. However,
considering potentially large scale everyday data, we can not feed all of them into the
rule engine. Rather, we need to be able to prune out most of the unnecessary metadata
except the ones potentially related to the desired reasoning task. Therefore, selecting an
appropriate list of facts from a bunch of surveillance metadata also requires understanding
semantics of each metadata items. Thus, to maximize the potential use of metadata, we
follow an ontological data representation and storing approach that captures dominant
concepts and relations employed in our surveillance system. The proposed data model
64
Contents
Six basic ontologies have been created. The first one is formed by data describing the
surrounding conditions and environment, the “Surveillance System Ontology”. The next
“Source Media Ontology” covers fundamental concepts and relations of possible input data
and sensors. The “Color Model” represents fundamental concepts and relations concerning
color model and is required for the similarity-based sub-query based on the MPEG-7 Color
Structure Descriptor [87]. For low-level features, we used slightly modified version of
“MPEG-7 Visual Descriptor Ontology” from AceMediaProject [156, 133]. While the “Video
Analytics” ontology models fundamental concepts and relations concerning the applied
video analysis algorithms, the “Domain Top Ontology” serves basic concepts and relations
Chapter 4
Figure 4.2 shows a partial description of these ontologies (For the ontology description
65
Contents
Figure 4.2: A Piece of the ‘Video Analytics’ Ontology Description in ORSD (Ontology
Requirement Specification Document) [163] Format.
66
Contents
Chapter 4
67
Contents
To extend logic programming with subjective logic, the CLIPS [1] rule engine was used as a
basis to provide flexibility for defining complex data structure as well as for providing a rule
resolving mechanism. To extend this system, we defined a data structure having the form of
6 with the following terms, agent (opinion owner), proposition, belief, disbelief, ignorance,
and atomicity. To represent propositions as first-order predicate logic (see Section 3.2.2 for
details), we extended the structure so that it can take arity n properties as well. Therefore,
given a predicate p the proposition can be described as ‘p(a1 , a2 , ..., an )’. In our system,
agent
therefore, each fact is represented as the form of ‘wp(a 1 ,a2 ,...,an )
’. Namely, rules are defined
with the opinion and predicate logic based proposition structure. Additionally, functions
of subjective logic operators taking opinions as parameters were defined. In this way,
uncertainty in the form of opinion triangle is attached to rules and facts. This aspect is
depicted as follows :
tive language and Subjective Opinion Space O, an opinion assignment over sentences k ∈ K
is a function ϕ : k → O. s.t
2. ϕRule : Rule → O, e.g. (wpacc(ac1 ,..,acn ) ← wpa11 (a11 ,..,a1n ) , .., wpain (ai1 ,..,ain ) ) = (b, d, i, a)
3. ϕRuleEval : RuleHead → ( a
⊛ wpaii(aai1 ,..,ain ) = O),
wpii ∈RuleBody
where ⊛ indicates one of subjective logic’s operators.
e.g. for a given rule wpacc(ac1 ,..,acn ) ← wpa11 (a11 ,..,a1n ) , .., wpain (ai1 ,..,ain ) ,
we interpret it as wpacc(ac1 ,..,acn ) = wpa11 (a11 ,..,a1n ) ⊛ ... ⊛ wpain (ai1 ,..,ain ) = (b, d, i, a)
It is important to note that there are different ways of opinion assignment. While
assigns an opinion to the consequence part of the rule (rule head). The assigned opinion is
68
Contents
functionally calculated out of opinions in the rule body using appropriate subjective logic
rules for default reasoning (see Chapter 5 and [82]). Given the initial opinion assignment
and an opinion assignment ϕ, labeling every sentence k ∈ K into Subjective Opinion Space
O, then the closure over k ∈ K, is the opinion assignment function cl(ϕ)(q) that labels
Chapter 4
For example, if ϕ labels sentences {a, b, c ← a, b} ∈ K as ϕf act (a), ϕf act (b) and
ϕRule (c ← a, b), then cl(ϕ) should also label c as it is information entailed by K. The as-
signment can be principled by the definition of closure. For example, an opinion assignment
to c, in a simple conjunctive sense can be ϕf act (a) · ϕf act (b) · ϕRule (c ← a, b), where · repre-
sent conjunction in Subjective Logic. In our system, to support the rich set subjective logic
that, most of rule based systems also support describing actions in the head part of a rule) :
Rule Body :
Due to the redundancy that arises when describing rules at the opinion structure level, we
wpacc(ac1 ,..,acn ) ← wpa11 (a11 ,..,a1n ) ⊛ ... ⊛ wpain (ai1 ,..,ain ) (4.2)
69
Contents
where ⊛ indicates one of subjective logic’s operators. This way of representing rules, we
can build a propositional rules that comprise opinions about a predicate as facts that
can also check logical conjunction based existence of involved opinions and finally define
resulted predicate with opinion attached by calculating opinion values with subjective logic
operators.
In this section, a case study is presented to apply the described system architecture to
a simple forensic retrieval scenario on assumption of wide area city surveillance. The
case study is designed to show the whole work-through and end-to-end work-pipeline of
the proposed system. In this case study, upon the manually annotated ontological data
“Who did this person talk to?” : Monitoring guards are watching live captured
video sequences in the monitoring room. They were informed that some event happened
around the area of camera 7. According to the report, a suspicious person wearing a
white T-shirt was seen was witnessed. While browsing the video archive of camera 7,
the guards found a person wearing a white T-shirt. They captured the frame, marked
the region of the person, and then narrowed down the location where the reported event
happened. By loading the query ontology, the system enables the “findTalkTo” query.
This query provides an automatic search for additional persons that had been talking
to the selected (suspected) person. By clicking the search button, the system starts to
search to resolve “TalkTo” relation for the selected person and reports a list of key frames
showing instances where the indicated person is talking to other persons. By Clicking
on the key frames the referenced video sequences are displayed on a monitor for a final
70
Contents
Chapter 4
Figure 4.4: Environmental Assumption : a) Locations of Nine Assumed Cameras, b) Some
of Frames from Camera 7 mounted in front of a Gas Station.
As an environmental assumption for the case study, ordinary video sequences are taken
during a day from nine different locations over an area. Figure 4.4 - a) shows the distributed
camera setup. Note that our test data has just normal and ordinary video sequences that
have no prominent unusual behaviors causing any alarm. For example, Figure 4.4 - b)
shows some of frames in our test data captured by camera 7, which is mounted in front of
a gas station.
71
Contents
In this first case study, we assume that all persons and all moving objects have been
along the proposed ontological metadata models. To take into account low-level features,
contents using our manual annotation tool is done by selecting bounding boxes from a given
video frames and designating a semantic concept defined in selected domain ontology. At
the same time, to incorporate low-level features and scene structure, we use a variation of
the MPEG-7 ontology (see Section 4.3.1 and [156, 133]). In order to represent the source
of the detection algorithm and related properties (such as, configuration parameters and
equipped camera the algorithm is running and so on) the tool allows additionally linking a
etc. Figure 4.5 shows a simple example of the annotation process with a semantically
labeled bounding box and the associated MPEG-7 CSD feature vector.
The generated annotation instances are encoded in RDF/OWL format. When RDF/
OWL technologies are used, the instances and owl concepts can be represented as DIG (Di-
72
Contents
Chapter 4
Figure 4.6: Annotation Instance of Figure 4.5 : a) DIG Representation of the Example, b)
Person Instance in N3-Triple Notation, c) ROI Instance in N3-Triple Notation
and d) Color Structure Descriptor in RDF/OWL Format of MPEG-7 Ontology.
rected Graph) consisting of RDF-triples of the form < subjective, predicate, objective >.
There are several encoding formats for RDF triples, among them N3-Triple (a.k.a Nota-
tion3) [169] is one of the simplest representation for human understanding. Figure 4.6
shows an example of a partial DIG of concepts and instances generated by the annota-
tion tool (ideally, this should be generated by the analytic modules automatically) and its
• the graph shows that there is an instance which is named as Person93 and having the
73
Contents
• The ROI exists in a video stream captured from camera id 4 (also defined in the system
profile).
• The ROI also has associated an instance of a visual descriptor named “CSD_inst_91”
Once generated, the metadata should be archived into a persistent repository for later
use. Because RDF/OWL is used as basis for the knowledge representation, it naturally
needs a RDF triple store as data archive. For practical use, the machine should be able
to parse such a graph model. There are several infrastructures available for this purpose
such as Sesame, Jena and Kaon, etc [175]. In the current system integration the Jena
semantic web framework from Hewlet-Packard (HP) is used, which is a Java framework
for building Semantic Web application. It provides a programmatic environment for RDF,
RDFS, OWL, and SPARQL (note. Jena is open source and grown out of work at the HP
labs).
indexed for picking similar vectors. As shown in Figure 4.6 - d), low level features usually
consist of a number of numeric values. In the above case, the shown CSD (Colour Structure
Descriptor) has 256 bins. Indexing such multi-dimensional vectors is however not a trivial
task. In textual data indexing as it has been done in RDBMS side, the most common way is
building a tree structure that balances the average degree of depth of nodes from the root,
by appropriately partitioning the search space. B-tree and its variations (B*-tree, etc) are
well known examples of this approach. However, it turns out that the partitioning is mean-
ingless when we deal with high dimensions, because no matter how we try to partition, the
74
Contents
Chapter 4
Figure 4.7: Illustrative Example of VA-file : a) VA-file for 2-D Vectors and b) Lower and
Upper Bound between a Given Query Vector q and Feature Vectors in a Cell.
curse of dimensionality causes the sparseness problem meaning that the data points in high
dimensional space will be dispersed so they cannot be well clustered. Multiple approaches
have been proposed so far: dimensionality reduction, vector approximation and locality
sensitive hashing technologies are well known examples [31, 129]. At this stage, considering
the simplicity on implementation, VA file [171] is used. VA file is based on the assumption
that we cannot partition data space that will in turn cause full data search. Therefore,
in the VA file approach, the focus is more on reducing I/O time and distance calculation
time. By quantizing each bin vectors with several bits, one specific feature vector can be
represented by a small number of bits. To cope with K-nearest neighbor search, it also
provides an algorithm to calculate distances at I/O time and thereby sort the requested
(bj ), as follows :
75
Contents
⌊ ⌋ 1 if j ≤ (b mod d)
b
bj = +
d
0 (4.3)
otherwise
This concept is shown in Figure 4.7 - a) in the case of b = 2 and d = 2. Once indexed,
the search can be done as shown in Figure 4.7 - b). While reading approximated feature
vectors, for each feature vector Pi we calculate Distmax (Ui ) and Distmin (li ) as shown
above. For K-nearest neighbour search, we first initialize an array with k-index with
appropriately large distance values, then with Pi we calculate the lower bound (Distmin )
if the lower bound is less than values in the array, we set the Pi as a candidate, and repeat
this process till we reach the end of the VA index file. Though the VA file approach is
described here, again, our ultimate aim is not to find another indexing structure, but to
provide knowledge on the index, so that the system can automatically decide what and
how to use such an index, when third party index structures are plugged into our system
As described, let us assume that someone informed that something happened somewhere.
With this information, we zoom into the reported area and click on one available camera
around the place. We also load a domain knowledge description, which will show us all
possible queries that we can resolve. Clicking a camera will show us a key frame summary
of what happened during the past period, sorted according to the time line. By clicking
on one of the images, we can play the corresponding video in the window on the top left
side (see Figure 4.8 - 1). While browsing the video sequences, we found a guy who is
holding a white box. We were suspicious about the guy and want to know as much as we
can (see Figure 4.8 - 2, 3). After capturing the scene (by clicking the capture button),
we start composing a query. This includes manually selecting the person by drawing a
bounding box, cropping the image and finally extracting a set of low level features for this
76
Contents
Chapter 4
Figure 4.8: Querying and Retrieval Graphical User Interface and Walkthrough.
region of interest (see Figure 4.8 - 4). From the available set of queries we select the query
‘FindTalkTo’ and trigger it (see Figure 4.8 - 5, 6). This in turn launches a SPARQL [168]
query that is expanded by the DL Reasoner using the Vision Metadata Ontology. The
• SELECT DISTINCT ?visionEntity ?am ?type ?ROIx ?ROIy ?ROI_width ?ROI_height ?camID
?timeStamp ?cvstr
#(opinion (agentl ?am) (prop ?type) (args ID ?visionEntity CamID ?camID RectInfo ”?ROIx ?ROIy
?ROI_width ?ROID_height Time ?timeStamp) #OpinionMapperMethod1(?cvstr)#)
FROM NAMED
<http://SicosleOntologyServer/metadata/KB/Sicosle/Instance/DataDocuments/MetaC1:040408:001.owl>
77
Contents
WHERE {
?visionEntity rdf:type sdtop:VisionEntity;
rdf:type ?type;
sdtop:hasRegion ?region;
vao:hasAnalyticModule ?am;
sdTop:hasConfidenceVale ?cvstr;
FILTER (regex (str(?type), ”ˆ http://SicosleOntologyServer/metadata/KB/Sicosle/DomainOnt/
Sicosle_Showcase1Domain.owl”)).
?region mpeg7:hasROITimeStamp ?timeStamp ;
mpeg7:ROI_x ?ROIx;
mpeg7:ROI_y ?ROIy;
mpeg7:ROI_width ?ROI_width;
mpeg7:ROI_height ?ROI_height;
sss:ROIOf ?videoStream .
# FILTER ((?timeStamp > ”2005-01-01 T00:00:00Z”ˆ ˆ xsd:dateTime)).
?videoStream ssm:capturedBy ?cam.
?cam rdf:type ssm:Camera .
?cam ssm:hasSensorID ?camID.
The result set is then transformed into a fact list for the CLIPS rule engine [1]. We
have also enabled describing hints on this translation in the comments (see the lines with
comment symbol ‘#’). Following shows examples of the translated opinion lists in CLIPS
• (opinion (agent humanAnnotator) (prop person) (args ID person96 CamID 4 RectInfo "259 75 58 136"
Time 2008-04-04T19:00:20.0) (b 1) (d 0) (i 0) (a 0.5) )
(opinion (agent humanAnnotator) (prop person) (args ID person93 CamID 4 RectInfo "159 55 50 126"
Time 2008-04-04T19:00:21.0) (b 1) (d 0) (i 0) (a 0.5) )
(opinion (agent humanAnnotator) (prop person) (args ID chair4 CamID 4 RectInfo "240 105 31 25"
Time 2008-04-04T19:00:21.5) (b 1) (d 0) (i 0) (a 0.5) )
(opinion (agent humanAnnotator) (prop person) (args ID table3 CamID 4 RectInfo "188 42 100 52"
Time 2008-04-04T19:00:19.0) (b 1) (d 0) (i 0) (a 0.5) )
:
As the metadata is based on manual annotation, the translated opinion values repre-
78
Contents
sent absolute truth (i.e., (1,0,0,0.5)). Once gathered, rules are executed on gathered data
segments, to find evidential patterns which satisfy given rules. Following shows a rule that
talkto_rule
wtalkto(append(V e2,V e3,T s2,T s3),CamID,(duration(T s2,T s3)),(dist(Roi2,Roi3)))
_ _ _
← wf urniture(V e1,CamID,Roi1,T s1) ⊛ wperson(V e2,CamID,Roi2,T s2) ⊛ wperson(V e3,CamID,Roi3,T s3) (4.4)
⊛wdistwithin(centerradius,100,Roi1,Roi2,Roi3)
SLSystem
⊛ wtimeholduntil(min,2,T
SLSystem
s1,T s2,T s3) ⊛ wneq(V e2,V e3)
SLSystem
Chapter 4
?op2<-(opinion(prop ?ty2&:(eq ?ty2 person))(args ID ?ve2 CamID ?camID RectInfo ?roi2 Time ?ts2))
?op3<-(opinion(prop ?ty3&:(eq ?ty3 person))(args ID ?ve3 CamID ?camID RectInfo ?roi3 Time ?ts3))
(distwithin centerradius 100 ?roi1 ?roi2 ?roi3)
(timeholduntil min 2 ?ts1 ?ts2 ?ts3)
(test (neq ?ve2 ?ve3))
=>
(bind ?new_op (sl-conjunction$ ?op1 ?op2 ?op3))
(modify ?new_op (prop talkto) (args ID (apply str-cat ”talkto” ?ve2 ?ve3 ?ts2 ?ts3)
CamID ?camID Time (duration ?ts2 ?ts3) Dist (dist ?roi2 ?roi3)))
)
Above rule also shows the use of subjective logic’s conjunction operator to derive
a new opinion. However, due to the manual annotation that gives full belief to every
metadata, the rule will also derive an opinion with full belief.
In this step it is e.g. verified in which instance at least two people are appearing
in the same image (because one of the rules defines that the presence of two persons is a
necessary condition for the ‘talk-to’ relation). The result from the rule based filtering will
be a largely reduced candidate set. The final check now is on verifying that one of the
persons is indeed the same as the initially selected (suspicious) person. This is performed
by a similarity match over all the features extracted from the selected ROI. If the distance
79
Contents
between the query feature set and the feature set associated to a person in the candidate
frame is below a certain threshold then it is assumed that both persons are identical. The
remaining candidate instances are presented as a list of key frames. Figure 4.8 - 7) shows
scenes with different people around chairs and table are regarded satisfying the given query.
Clicking ‘view info menu’, it also shows visual similarity matching results using L1 − norm
between features of objects in the retrieved scenes and the extracted query feature vector.
In addition to this, ideally, these retrieved items can be clicked to check the attached
subjective opinion on being the right answer to the query Figure 4.8 - 8). However, at this
stage of case study, we have not considered the uncertainty handling mechanism but rather
used straightforward rule matching. The detailed mechanism on uncertainty handling will
In this section, we present a case study in applying the proposed system to real traffic
surveillance scenes processed with automated vision analytics 1 . Considering that we are
not yet facing the use of vision analytic metadata generated from heterogeneous vision
analytic systems or vendors, we will focus more on logical reasoning aspect rather than the
metadata representation (note that, one of the reasons we use ontological metadata model
giving a sharable formal semantics to metadata. Refer to the Section 2.2.3, Section 4.2 and
Section 4.3.1). Therefore, this case study is to show how logical rules can be used to model
contextual semantics implicitly implied in traffic video scenes. For the convenient imple-
mentation, the generated metadata is directly archived into a DataBase (MySQL) in form
1
This case study is conducted in cooperation with SCR (Siemens Corporate Research, Princeton, NJ,
US) using SCR’s proprietary vision analytic modules. Especially, special credits must go to Dr. Vinay
Shet, Dr. Gao Xing and Dr. Vasu Parameswaransu
80
Contents
Figure 4.9: Samples of Traffic Visual Surveillance Data for Case Study 2.
Chapter 4
4.5.1 Scenario Setting for Case Study II
Total 5 hours and 30 minutes ordinary tunnel and highway scenes are collected from 8
different video sources. These video sources are processed using SCR’s (Siemens Corpo-
rate Research) proprietary automated vision analytic modules. The volume of the video
footages is 4.64Gb in compressed form. Figure 4.9 shows some sample frames of the traffic
video data. The applied vision analytic modules generate total 12 primitive semantics (e.g.
trajectory information of vehicles, distances between vehicles, scene with hazard light, oc-
clusion, human appear and disappear, etc.). Upon such semantics, we focus on composing
and answering complex queries by the use of spatio-temporal rules. Following shows the
81
Contents
Table 4.1: Metadata Vocabulary and the Property of Each Primitive Semantic Item.
Figure 4.10: Flat-file Style Database Tables in ERD (Entity Relation Diagram) Each Rep-
resenting Metadata Vocabularies of Vision Analytics.
Table 4.1 shows the metadata vocabulary used in this case study. Using the vision analytics
available in Siemens (i.e., SCR’s proprietary vision analytics), total 12 primitive semantic
items are generated with corresponding properties. Archiving of the metadata is done using
82
Contents
a flat-file-like database scheme shown in Figure 4.13, which directly maps each semantic
the [Ef, Ea] pair that satisfies Ef + Ea = 1. The value Ef represents amount of truth
‘Evidence For’ and the other value Ea represents amount of falsity ‘Evidence Against’.
In our system, each metadata item is represented in form of subjective opinion following
SCR_V isionAnalytic
Definition 16 - 1, as shown in Section 4.3.2. For example, whumanappear(CamID,P edID,T ime) =
Chapter 4
To be able to answer complex queries Q1-Q7 discussed in Section 4.5.1, it should be able
to augment implicit high level semantics, based on the currently available metadata. For
example, the semantics of ‘emergency vehicles’ in ‘Q1’ is not explicitly defined in the meta-
gency vehicle’ if there was also a hazard light at the same time of the vehicle detection.
Using the logical conjunction operator (‘∧’) of subjective logic, this can be represented as
follows :
evrule
wemergencyV ehicle(CamID,V ehID,V ht)
_ _
← whazardlight(CamID,_,Ht) ∧ wvehicletrajectory(CamID,V ehID,V ht) (4.5)
∧wtest(V
SLSystem
ht+25<Ht<V ht+25)
The semantic concept ‘truck’ in ‘Q1’ is also not explicitly defined in the metadata
vocabulary. However, we can also infer the concept ‘truck’ by referring to ‘vehicletype’ that
indicates a large vehicle when the property ‘type’ is set with ‘1’. This can be also depicted
as follows:
truckrule
wtruckV ehicle(CamID,V ehID,V ht)
(4.6)
_ _
← wvehicletype(CamID,V ehID,1) ∧ wvehicletrajectory(CamID,V ehID,V ht)
83
Contents
Based on the inferred concept defined by Rule 4.5 and Rule 4.6, the semantics of
q1
wtruckN earAnEmergencyV ehicle(CamID,T vID,EmerT )
_ _
← wemergencyV ehicle(CamID,EvID,EmerT ) ∧ wtruckV ehicle(CamID,T vID,V ht) (4.7)
∧wappearEqual(EmerT,V
SLSystem
ht,70)
Rule 4.7 represents that it will assert a new opinion on the proposition ‘truckNear-
and ‘truck vehicle’ at the same time. To manipulate temporal relations, Allen’s temporal
relations were modeled as predicate function [15]. These temporal relation covers ‘{equal,
before, after, meets, meetby, overlaps, overlappedby, starts, startedby, during, contains, fin-
ishes, finishedby }. For example, the predicate function term ‘appearEqual’ corresponds to
‘equal’ in allen’s temporal logic with acceptable tolerance 0.7 sec. Semantics of ‘Q2-Q7’
can be also augmented in the similar way. Table 4.2 shows some of augmented semantics
in form of rules and all the rules directly corresponding to the queries ‘Q1-Q7’.
Figure 4.11 shows the prototype interface of the demonstrator. The user interface
shows retrieved items for the queries ‘Q1-Q7’. Each items comes with basic information and
the item can be played for further examination. The items are ranked along the expectation
value of the calculated subjective opinions attached to items. In this case study, however,
the uncertainty handling is only based on the ‘conjunction’ operator, thereby, represent
rather conjunctive probability sense of subjective opinions. (Note that, the rest of this
dissertation deal with the detailed aspects on handling uncertainty and this section focusses
more on the architectural case study.) As an architectural proof of concept, every queries
seem to retrieve reasonable scenes that matches to the intention of the queries. However,
for some cases as shown in Figure 4.11 - ‘Q4 : Find a truck going past a Pedestrian’, due to
the false alarms of pedestrian metadata, it also retrieved some of wrong items. Therefore,
84
Contents
evrule
wemergencyV ehicle(CamID,V ehID,V ht)
_ _ SLSystem
← whazardlight(CamID,_,Ht) ∧ wvehicletrajectory(CamID,V ehID,V ht) ∧ wtest(V ht+25<Ht<V ht+25)
(4.5)
truckrule
wtruckV
_ ehicle(CamID,V ehID,V ht) _ (4.6)
← wvehicletype(CamID,V ehID,1) ∧ wvehicletrajectory(CamID,V ehID,V ht)
bvrule _
wbackingU pV ehicle(CamID,V hID,V ht) ← wspeedestimation(CamID,V hID,1,V ht) (4.8)
:
‘Q1’ Find a Truck following an Emergency Vehicles
q1
wtruckN earAnEmergencyV ehicle(CamID,T vID,EmerT )
_ _
← wemergencyV ehicle(CamID,EvID,EmerT ) ∧ wtruckV ehicle(CamID,T vID,V ht) (4.7)
SLSystem
∧wappearEqual(EmerT ,V ht,70)
Chapter 4
‘Q2’ Find an Emergency Vehicle following another Emergency Vehicle
q2
wemerV ehicleF ollowingAnotherEmerV ehicle(CamID,EvID2,EmT 2)
_ _
← wemergencyV ehicle(CamID,EvID1,EmT 1) ∧ wemergencyV ehicle(CamID,EvID2,EmT 2) (4.9)
SLSystem
∧wappearBef ore(EmT 1,EmT 2,80)
q3
wvehicleGoingP astP edestrian(CamID,V hID,Ht)
_ _ SLSystem
← whumanappear(CamID,P edID,Ht) ∧ wvehicletrajectory(CamID,V hID,V t) ∧ wappearBef ore(Ht,V t,80)
(4.10)
q4
wtruckGoingP astP edestrian(CamID,V hID,Ht)
_ _ SLSystem
← whumanappear(CamID,P edID,Ht) ∧ wtruckV ehicle(CamID,V hID,V t) ∧ wappearBef ore(Ht,V t,80)
(4.11)
q5
wtruckP assingByStoppedV ehicle(CamID,V hID,St)
_ _ SLSystem
← wstoppedvehicle(CamID,_,_,St) ∧ wtruckV ehicle(CamID,V hID,V t) ∧ wappearBef ore(St,V t,80)
(4.12)
85
Contents
Figure 4.11: Querying and Retrieval Graphical User Interface and Retrieval Results of ‘Q1-
Q7’ with Sample Items.
86
Contents
4.6 Discussion
In the previous section, we have demonstrated the proposed architecture approach in two
different case studies. These case studies especially show the feasibility of logical reasoning
for modeling compositional high-level semantics that can be augmented based on available
metadata. While the reasoning aspect is focussed, there are however still open issues such as
‘Scalability’, ‘Proper Uncertainty Handling’, etc. In this section we will further discuss the
‘Scalability’ issue and the issues around ‘Uncertainty Handling’ will be addressed in the rest
of this dissertation. In the proposed architectural pipeline, there are two main components
Chapter 4
handling’ and the ‘scalability issues around rule-based reasoning’. This section addresses
proper selection of metadata segments for further logical reasoning. The selection is done
by the use of SPARQL [168] query upon OWL/DL reasoner [175]. Although there are
issues such as how to come up with Ontologies that can be agreed by other systems and
venders, etc. (‘in the sense of knowledge sharing and interoperability’), the more essential
and natural question here could be the performance in terms of the scalability due to the
potentially large scale metadata that visual surveillance systems would produce. It is not
only the case of visual surveillance but also the case of many other domains. Therefore, it
There have been performance benchmarks of currently available reasoners [67, 172,
173, 30]. The benchmarks have been conducted in terms of many aspects such as size
loading time of Ontology, complexity of the queries and currently available large scale
ontologies (i.e., DBLP database for publications, etc.). In a relatively recent comprehensive
87
Contents
Table 4.3: Statistics of Test Ontologies used in the Work of Jürgen Bock et al. [30]. The
Table Indicates # of TBox Parameters (Class, Property, SubClass, etc.) and #
of ABox Parameters (C(a) : Instances of a Class and Instances of a Relation
R(a,b)). The Red Boxed Row Shows the Statistics of Our Manual Annotation
Used in Case Study I (see Section 4.4) Extracted by TopBraid Composer. ‘-’
Represents the Statistics not Available in TopBraid Composer. The Red Ar-
row Indicates that Our Annotation Resembles wine_1 Data Set in terms of
Statistics.
workbench report done by Jürgen Bock et al. [30], they have conducted a performance
Table 4.3 shows the statistics of the test ontologies. We have also extracted the
statistic information of our annotation using an ontology editing tool called ‘TopBraid
Composer’ 3 . Although not every statistics were available as indicated with ‘-’, the red
2
Datasets: VICODI (http://www.vicodi.org), SWRC (http://www.ontoware.org/index.html),
LUBM (http://swat.cse.lehigh.edu/projects/lubm/index.htm) and WINE (http://www.
schemaweb.info/schema/SchemaDetails.aspx?id=62)
3
http://www.topquadrant.com/products/TB_Composer.html
88
Contents
Chapter 4
Figure 4.12: Benchmark Result in the Work of Jürgen Bock et al. [30]. The Figures show
the Average Load and Query Time with a Particular Reasoner on the Dataset
shown in Table 4.3. The ‘◦’ indicates Time-out (> 5 min).
box and the red arrow in Table 4.3 shows that our annotation is quite similar to the
wine_1 data set. Therefore, it is expected that our ontology model would also show
rently available. The reasoners used in their evaluations fall into three groups: tableau
algorithm based (RacerPro and Pellet), disjunctive datalog based (KAON2) and standard
89
Contents
Figure 4.12 shows the result of performance evaluation. In the benchmark report,
they informed that it was conducted on a Linux 2.6.16.1 system, however, no further
information regarding the hardware setting is available. Unfortunately, it shows that many
of reasoners except KAON2 becomes ‘time-out’ when it is performed with wine data sets
containing more than 5000 instances. In the case of wine_1 data set on Pellet, as expected,
it shows similar performance to the result we experienced with our annotation in the case
study I (Including loading and query processing time, it took approximately 20-25 sec on a
core2duo T7200 1.8 Ghz, 2Gb Ram Windows XP machine). According to the benchmark,
it does not come up with a clear ‘winner’ OWL/DL reasoner that performs well for all
types of ontologies and reasoning tasks. However, Jürgen Bock et al. [30] have concluded
as follows:
• As general conclusions we can summarize our results in that (1) reasoners that employ
a simple rule engine scale very well for large ABoxes, but are in principle very limited
to ‘lightweight language fragments’, (2) classical tableau reasoners scale well for complex
TBox reasoning tasks, but are limited with respect to their support for large ABoxes, and
in KAON2 scale well for large ABoxes, while at the same time they support are rich
language fragment. If nominals are important for a given scenario, Pellet is the only
The benchmark shows that the complexity of TBox also influence the ABox perfor-
mance. To this extent, the word ‘lightweight language fragments’ in (1) means OWL-Lite
tractability. At this stage of the proof of concept, we have used Pellet reasoner not to be
restricted by expressivity on knowledge modeling. However, it seems that a lot more effort
should be paid to derive a matured ontological knowledge model in visual surveillance do-
main, that is not only agreeable by the parties of systems or venders but also balanced in
pellet/index.shtml), KAON2(http://kaon2.semanticweb.org/), Sesame(http://openrdf.org)
and OWLIM(http://www.ontotext.com/owlim)
90
Contents
terms of TBox and ABox complexity. Therefore, we believe that this should be driven in
the sense of ‘collaborative standardization’. Both the ‘scalability’ and the ‘standardization
of ontology model’ hamper the wide acceptance of the ontology related technology not only
in the visual surveillance but also in other domains. This seems the reason why many of
current data management systems remain in the traditional database technology. However,
considering the current active research effort paid in OWL related technologies, we believe
that it would come up with improved performance in the near future. To this extent, we
believe that the proof of concept (case study I) on the use of ontological metadata shows
Chapter 4
4.6.2 Scalability of Rule Engines
archive information in a ‘sharable’ and ‘machine interpretable’ way, but also to provide
a mechanism to pick and collect probable amount of partial evidences (called facts in
another important focus regarding ‘scalability’ is on the rule based logical reasoning engines
that are used after collecting evidential facts. Normally, rule engines implement traditional
backtracking (e.g., SLD resolution [66]) or forward chaining (e.g., Rete algorithm [64])
algorithms. In such algorithms, facts and rules are loaded in memory space called fact
base and rule base respectively. It is known that the size of the facts affects reasoning
performance.
bility’ of representative rule engines such as CLIPS [1] and Prolog 5 . Although it is not
designed for ‘automated’ inference, in logical view, it is important to note that the ‘re-
lational data model’ is also based on the ‘first-order-predicate-logic’ [40, 41] and SQL is
the language to access the relational data model. Namely, they operate across the same
5
For the evaluation, we used SWI-Prolog [8] among many variations due to its GNU-license policy. Com-
parison of Prolog implementations can be found on (http://en.wikipedia.org/wiki/Comparison_of_
Prolog_implementations)
91
Contents
Figure 4.13: The Database Scheme of a Siemens’ Internal Visual Surveillance Prototype
System called VAP (Video Analytic Platform).
conceptual theory, although their focuses are in completely different directions. In rule
engine terms, SQL is primarily a fact and relation(set) engine therefore does not support
automatic pattern tracking that dynamically builds a rule tree at run time. However, giv-
ing up the ‘automatic’ inference aspect, and by taking static view of all possible queries,
SQL can do the other, to a limited extent. Therefore, to contrast the performance to the
case of using traditional database, we manually expanded the chaining of the rules for each
queries ‘Q1-Q7’ in form of SQL. To see the influence of different relational data models,
the SQL expansion is done for two different relational models. The first one is the flat-file
style model as shown in Figure 4.13 explained in Section 4.5. We have also used a more
complex data model that covers many other aspects of surveillance system. The model is
shown in Figure 4.13 and is used in one of Siemens’ internal prototype implementations of
surveillance system called VAP(Video Analytic Platform). Table 4.4 shows some of SQL
query examples expanded upon the two different relational data models.
We have also populated instances of the metadata vocabulary set used for case study
II in Section 4.5. As shown in Table 4.1, originally, 40094 instances were extracted from 8
92
Contents
select distinct hl.CamID, hl.Time, v.time select distinct ep1.camid, ep1.time, ep2.time
from hazardlight hl, vehicletype vt, vehicletrajectory v from event_property ep1, event_property ep2
where hl.CamID = vt.CamID where ep1.eventtypeid = 2
and vt.CamID = v.CamID and ep1.proptypeid = 7
and vt.VehicleID = v.VehicleID and ep2.eventtypeid = 13
and vt.Type = 1 and ep2.proptypeid = 9
and v.Time > hl.Time and ep2.propvalue = 1
and hl.Time + 80 > v.Time and ep1.camid = ep2.camid
(4.15) and ep2.time > ep1.time
and ep1.time + 80 > ep2.time
(4.16)
select distinct v.CamID, v.Time, d.time select distinct ep2.camid, ep2.time, ep1.time
from distance d, vehicletrajectory v, from event_property ep1, event_property ep2,
vehicletype vt, vehicletype vt2 event_property ep3, .event_property ep4,
where vt.Type = 1 event_property ep5, event_property ep6,
and vt.camID != 0 event_property ep7
and v.CamID = vt.CamID where ep2.eventtypeid = 13
and v.VehicleID = vt.VehicleID (4.17) and ep2.camid != 0
Chapter 4
and d.CamID = v.CamID and ep2.proptypeid = 9
and v.VehicleID = d.VehicleID and ep2.propvalue = 1
and d.Distance < 25 and ep1.eventtypeid = 1
and vt2.Type = 1 and ep1.camid = ep2.camid
and vt2.CamID = vt.CamID and ep3.id = ep1.id
and d.AssocVehicleID = vt2.VehicleID and ep3.eventtypeid = 1
and ep3.proptypeid = 1
and ep4.id = ep2.id
and ep4.eventtypeid = 13
and ep4.proptypeid = 1
and ep3.propvalue = ep4.propvalue
and ep1.proptypeid = 6
and ep1.propvalue < 25
and ep5.eventtypeid = 13
and ep5.proptypeid = 9
and ep5.propvalue = 1
and ep2.camid = ep5.camid
and ep6.id = ep5.id
and ep6.eventtypeid = 13
and ep6.proptypeid = 1
and ep7.id = ep1.id
and ep7.eventtypeid = 1
and ep7.proptypeid = 2
and ep6.propvalue = ep7.propvalue
(4.18)
Data Set # instances time of each video sources total time of 8 video sources
Original Case Study 2 # 40094 ≒ 40 min ≒ 5 hours 30 min
CS2-0.5 27290 ≒ 30 min ≒ 4 hours
CS2-01 42444 ≒ 1 hours ≒ 8 hours
CS2-02 84888 ≒ 2 hours ≒ 16 hours
CS2-03 127332 ≒ 3 hours ≒ 24 hours
CS2-04 169776 ≒ 4 hours ≒ 32 hours
CS2-05 212220 ≒ 5 hours ≒ 40 hours
CS2-06 254664 ≒ 6 hours ≒ 48 hours
CS2-07 297108 ≒ 7 hours ≒ 56 hours
CS2-08 339552 ≒ 8 hours ≒ 64 hours
CS2-24 1018195 ≒ 24 hours ≒192 hours
CS2-48 2036390 ≒ 48 hours ≒ 384 hours
Table 4.5: Statistics of Dataset prepared for the Scalability Evaluation of Rule-engines and
Relational Database.
93
Contents
different video sources. The total volume of the data is about 5 hours and 30 minutes, that
is about 40 minutes per each source. Based on the original data set, we have populated
dummy instances to prepared data sets as shown in Table 4.5. The data sets are designed
to contain number of dummy instances correspond to varying hours. For example, the data
set CS2-48 contains, 2036390 instances that corresponds to 384 hours (48hours / source ×
8 source).
94
Contents
For the evaluation, we set SWI-Prolog 5.6.51, CLIPS 6.3 and MySQL 5.0 on a
core2duo T7200 1.8 Ghz, 2Gb Ram Windows XP machine. Figure 4.14 shows the eval-
uation results. Overall, SQL performs better than rule engines as expected. However,
this was not always the case that SQL performs better than rule engines. For example,
in the case of ‘Q2’, the VAP scheme showed worst result. Indeed, the cases ‘Q1’, ‘Q3’
and ‘Q7’ show similar results between CLIPS and VAP Scheme. This seems because the
relational models (schemes) were not optimized as it is usually done in the reality of appli-
cation development. Therefore, it seems that there would be some optimization issues such
as setting proper indices and separating tables according to correlated data density, etc.
Overall, CLIPS performed better than Prolog. The rules can also be optimized considering
Chapter 4
the data distributions so that it can traverse less amount of instances first in the case of
conjunctive rules. Therefore, speaking overall assessment on the evaluation, it seems that
rule engines could also cover about 150,000 to 200,000 instances of data (in total about 32
to 40 hours 8 camera video sources in our setting) within 5 minutes response time, when
the rule optimization is also considered. Nevertheless, traditional relational database seems
still promising in terms of response time although they can not support automated infer-
structures. Therefore, as a hybrid approach, one could consider the automated rules to
single SQL query mapping mechanism, that can be triggered once rule engine binds rule
search tree. To this extent, it appears many literatures on this issue. Some of them are
to simply interface rule engine with relational database [39, 106, 112], to translate and to
compile rules in prolog to SQL [55] and to optimize rules relational query system [89].
In summary of lessons learned in this chapter, while the attempt of using ontological
representing and resolving epistemic contextual semantics, the scalability is not sufficient
95
Contents
lance systems with high potential usage. Reminding the requirements discussed in Sec-
tion 1.2.4, such efficient semantic analysis of visual surveillance data requires a system
oriented approach that optimally combines the individual legacy vision analytic power and
its optimal use. In this chapter, we discussed the software components, data processing
pipeline and system architecture design. Among others, the main design philosophy of
our approach was to maximize utilization of available analytic metadata together with
context information. In this view, the proper vision analytic metadata representation and
the contextual knowledge expressive power of the reasoning framework are critical issues
to accomplish the goal. For the former requirement, we introduced ontological metadata
representation and for the latter requirement, we adopted logic programming. The main
based ontological data representation model (this is due to the ‘modularity’ as explained
in Section 2.3.3).
While the reasoning capability can be applied both the real-time analysis and the
forensic post analysis, the two case studies show the potential and the feasibility of the
practical real applications, the reasoning power should be ‘scalable’. Section 4.6 shows
performance benchmark of triple stores and rule-engines against different scale of metadata
(i.e., fact base in view of rule-engines). Unfortunately, it seems that logical formalism based
approaches and the triple stores of ontological metadata are by themselves not sufficiently
scalable. The ‘scalability’ problem is common issue in the realm of ontology related research
field and also expert system related researches. One good news, however, is that there have
been undergoing active research focus on the ‘scalability’ issue. We have briefly introduced
some possible remedies shown in literatures. In addition, the ‘scalability’ of low-level feature
96
Contents
matching is also important and is an active research topic in the realm of high dimensional
vector indexing field. Therefore, we believe and hope that we could benefit from those
Another critical and rather essential issue is the proper handling of ‘uncertain’ nature
of vision analytic metadata and knowledge models. In the rest of this dissertation we will
Chapter 4
97
98
5 Contradictory Information Handling
Thus far, we have discussed architectural aspect of the proposed approach to high-level se-
mantic analysis of visual surveillance data. The main advantage of the proposed approach
is on the high degree of ‘modularity’ that makes more ‘flexible’ contextual knowledge mod-
eling possible. Besides the ability of flexible context modeling, ‘uncertainty’ handling
mechanism is also an important issue. While we have briefly explained the extension of
logic programming with subjective logic (see Section 4.3.2), the detailed aspect to sup-
port ‘epistemic reasoning’ under ‘uncertainty’ is not covered in the previous Section. In
Chapter 5
this chapter, we further deal with the ‘uncertainty’ aspect starting with the discussion
on the ‘default reasoning’ that supports ‘non monotonicity’ explained in the preliminaries
Section 5.5.
5.1 Introduction
In forensic analysis of visual surveillance data, ‘default reasoning’ can play an important
role for deriving plausible semantic conclusions under ‘imprecise’, ‘incomplete’, ‘conflict’
and ‘contradictory’ information about scenes. In default reasoning, not only facts (analytic
metadata) but also rules (contextual knowledge) are the sources of ‘imprecise’ and ‘con-
the proper representation of ‘epistemic belief strength’ about given piece of information
but also the proper ‘principled fusion’ of the information is required. A discrete species of
Bilattice for multivalued default logic is one that demonstrated default reasoning in visual
99
Contents
logic that acts in a continuous space. As an uncertainty representation and handling for-
malism, subjective logic bridges Dempster Shafer belief theory and second order Bayesian,
thereby making it attractive tool for artificial reasoning. For the verification of the pro-
posed approach, we further extend the inference scheme on the bilattice for multivalued
default logic to L-fuzzy set based logics that can be modeled with continuous species of bi-
lattice structures. We present some illustrative case studies in visual surveillance scenarios
to contrast the proposed approach with such L-fuzzy set based approaches.
The main objectives of this chapter is as follows: 1) to bestow ‘default reasoning’ capa-
bility upon our subjective logic extended logic programming framework, 2) to compare
the proposed framework with other uncertainty formalisms that could alternatively model
Section 2.3. Extensional approaches treat knowledge as conditional rules that are labeled
with uncertainty [130]. For the labeling uncertainty, in this dissertation, we proposed
the use of subjective logic [93]. However, when it comes to uncertainty representation
formalisms, there are number of other formalisms such as Bilattice [69], fuzzy set based
fuzzy logic [176, 177], Dempster Shafer belief theory [149] and traditional probability based
Bayesian approaches [33], etc. Therefore, for the proper choice of uncertainty formalism,
handling uncertainty.
As explained in Section 3.1, subjective logic [92, 93] is also one such uncertainty rep-
resentation and handling formalism that can be seen as extended theory derived from both
the Dempster Shafer belief theory and the second order Bayesian. From Dempster Shafer
belief theory, subjective logic inherits the philosophy of explicit representation of ignorance
100
Contents
about knowledge in a model called subjective opinion triangle that can be also mapped
into beta distribution. The operators of subjective logic are also derived in the sense of
Bayesian. Unlike traditional Dempster Shafer evidence fusion method, that is known to
yield counter intuitive result when it is operated with highly contradictory evidences and
also known to be inconsistent with Bayes’ rule, subjective logic comes with similar opinion
fusion operators that are robust even with such highly contradictory evidences [92]. Com-
pared with bilattice that mainly consists of two lattices, one representing degree of truth
and the other representing degree of information respectively, the degree of information
concept is similar to the degree of ignorance in subjective opinion. The main difference
between bilattice and subjective logic is the operators. While bilattice comes with four
operators that are compositionally defined based on two lattice operators meet and join
from the perspective of set theory, subjective logic comes with 12 operators defined rather
in Bayesian sense. Another formidable uncertainty handling formalism, fuzzy logic is based
on fuzzy set theory that relies on degree of membership concept for a knowledge segment
Chapter 5
and again this is similar to the concept of partial ignorance in subjective logic. Interest-
ingly, it is known that some extensions of fuzzy logics can be modeled with (bi-)lattice
structures. One thing worth to note concerning fuzzy logic is that, even though there are
Zadeh’s original logical operators, there are yet another ways of defining logical operators
as well. However, due to this aspect, there is inconsistent between fuzzy logic operators and
classical probability calculus, thereby often criticized by statisticians who prefer Bayesian
[179]. Thus, we advocate that above aspects make the use of subjective logic attractive as
of a partial knowledge segment is often easy to be fragile, because there can be potentially
possible contradictions or counter examples about the given knowledge segment. Due to
this aspect, the property of retracting and updating existing beliefs upon acquisition of
new information (aka. belief revision) is essential. Default reasoning introduced by Reiter
101
Contents
[143] is one such non-monotonic reasoning method especially under contradictory know-
ledge segments. Default reasoning allows expressing a segment of knowledge as being ‘true
by default’ or ‘generally true’, but could be proven false upon arrival of new information.
one that is used to model default reasoning and demonstrated the usefulness on performing
human identity maintenance and contextual reasoning of event in visual surveillance do-
main [153, 154]. As noted above, the degree of truth and the degree of information concepts
in bilattice are similar to the ones in subjective logic. Focusing on the similarity, we exam-
ine subjective logic operators that have corresponding semantic behavior to the operators
defined on bilattice framework. As mentioned above, what is also interesting is that some
continuous species of bilattice structures are often used to represent two different species
of fuzzy logic. Namely, intuitionistic (or interval-valued) fuzzy logic that can be modeled
with so-called ‘triangle’ bilattice and fuzzy Belnap logic (i.e., fuzzified four-valued logic,
FOUR) that can be also modeled with so-called ‘square’ bilattice [18]. The relationship
between these two fuzzy species of bilattice structures is studied in the work of Cornelis
et al. [42, 43] and Arieli et al. [18, 19]. Interestingly, the uncertainty representation in
intuitionistic fuzzy logic (‘triangle’) is very similar to that of the opinion triangle. There-
fore, to verify the proposed subjective logic based default reasoning approach and to study
its similarity and dissimilarity with fuzzy logics, we further extend the inference scheme
defined on the discrete bilattice structure for the multivalued default logic onto the two
continuous species of bilattice structures. To better verify and contrast the characteristics
of the proposed approach, we present some illustrative case study examples in typical vi-
sual surveillance scenarios. We then compare the default reasoning results yielded from
the proposed subjective logic based approach, bialttice for multivalued default logic, the
intuitionistic fuzzy logic (‘triangle’) and the fuzzy Belnap logic (‘square’ ).
We believe this way of comparison better position the subjective logic as a tool for
artificial reasoning and also give us better insights on the correlations among different
102
Contents
bridge between Dempster Shafer belief theory and Bayesian. Then by the comparison in
this work on modeling default reasoning, it shows the bridge among subjective logic, fuzzy
is becoming more important. To address this aspect, there has been some work on the
use of logic programming languages due to the expressive power and on the use of differ-
uncertainty as a generalized truth value attached to formulas and compute the uncertainty
of any formula as a function of the uncertainties of its sub formulas [130]. Akdemir et
Chapter 5
al. [13] used an ontology structure for activity analysis, but with no uncertainty handling
mechanism. Shet et al. [152] introduced a system that adopts Prolog based logic program-
ming for higher-level situation reasoning in visual surveillance. The same authors adopted
bilattice based multivalued default reasoning for identity maintenance of human detection
results and context reasoning [153, 154]. Jianbing et al. [110] adopted Dempster Shafer
belief theory with the use of rule-based system for bus surveillance scenario. Anderson et
al. [17] adopted fuzzy logic to model and analyze human activity for video based eldercare
scenario. While different uncertainty handling formalisms are introduced with logic pro-
gramming based knowledge modeling, principled handling of default reasoning has been
only demonstrated by the bilattice based approach [153, 154] (refer to Section 2.4.3 for
When it comes to bilattice framework itself, it is known that some continuous species
of bilattice structures that are called ‘triangle’ and ‘square’ correspond to intuitionistic
fuzzy logic and fuzzy Belnap logic, respectively [18]. Naturally, there has been comparative
103
Contents
study on the characteristics between intuitionistic fuzzy logic and fuzzy Belnap logic [42,
43]. Atanassov [22] introduced a transformation between these two fuzzy logics and proved
that the transformation is bijective. The use of square bilattice is demonstrated to improve
human detection results by the use of rule-based reasoning given high false alarm rate and
partial occlusion based output of different body parts based detectors [155] with the similar
inference scheme shown in their previous work [153, 154]. In this chapter, we show that
the default reasoning behavior on multivalued default and square bilattice can be also
modeled using subjective logic. Relying on the study of Atanassov [22] and Cornelis et al.
[42, 43], we also show the correspondence among subjective logic, intuitionistic fuzzy logic
5.4 Preliminaries
This section gives an overview of the fundamental background about uncertainty represen-
tation and handling formalisms that will be discussed in this chapter in the view of default
reasoning. The preliminaries will cover bilattice theory, and two extensions of fuzzy log-
ics, namely, intuitionistic fuzzy logic and fuzzy Belnap logic. Refer to Chapter 3 for the
Bilattices are algebraic structures which are mainly built on top of the concept poset (i.e.,
a property of our interest) introduced by Ginsberg [69] and elaborated by Fitting [60].
Definition 18. (Partial Order) . A partial order is a binary relation ≤ over a set S
which is reflexive, antisymmetric and transitive, i.e., for all a,b and c in S, satisfies : a)
then a ≤ c (transitive).
104
Contents
Definition 19. (Poset) . A set S with a partial order (S, ≤) is called partially ordered
a) for any two elements a and b of L, the set {a, b} has a least upper bound ∨ (join).
b) for any two elements a and b of L, the set {a, b} has a greatest lower bound ∧ (meet).
B = (B, ≤t , ≤k ), where B is a nonempty set containing at least two elements, and (B, ≤t ),
(B, ≤k ) are complete lattices (for which all subsets are also lattices).
Chapter 5
The name ‘bi’ - lattice indicates that it is a structure consists of two lattices. Lattices
are any poset that are possible to define meet ∧ (aka. greatest lower bound) and join ∨
(aka. least upper bound) for any two elements in it. A partial order is a generalization
of ordering, i.e., a binary relation ≤ over a set S which is reflexive, antisymmetric and
transitive. Lattices are often expressed as a graph whose edges represent the binary relation
of ‘partial order’ between two elements that can be directly compared. There can be
elements a and b in the lattice L for which an order between them cannot be determined.
However, greatest lower bound meet (a ∧ b) and the lowest upper bound join (a ∨ b)
can always be determined for any of two elements a and b in the lattice L. Namely, by
considering two sets that contain elements that are greater than a and greater than b
respectively, the lowest element which can be found in both of the sets is the meet (and
join can be similarly defined). Figure 5.1 (a) shows a lattice that some elements in it, for
in lattice theory, ‘1’ represents the greatest elements and ‘0’ is the lowest elements and
105
Contents
Figure 5.1: Examples of Lattice and Bilattices (a) a Lattice (b) Bilattice corresponding to
Traditional Binary Logic (c) Bilattice corresponding to Three-valued Logic (d)
Bilattice corresponding to Belnaps Four-Valued Logic, FOUR. (Refer to [69]
for more detail).
therefore for any a in lattice L, a∧1 = a, a∨1 = 1, a∧0 = 0 and a∨0 = a. Bilattices provide
semantics for reasoning by considering one lattice with partial order in terms of degree of
truth ≤t and the other lattice with partial order in terms of degree of information ≤k (note
that, the semantics of degree of information often can be seen as degree of specificity of
To avoid the confusion that would arise from using the same symbols meet ∧ and
join ∨ for both lattices, following Fitting, we use the symbols meet ⊗ and join ⊕ for
the lattice with partial order ≤k [60]. While the meaning of ∧ and ∨ corresponds to
the standard logical role of conjunction and disjunction, the meaning of ⊗ and ⊕ are
less intuitive. Fitting named ⊗ as consensus operator in the sense that it derives the
most degree of information agreed upon two operands [60, 61]. Likewise ⊕ is named as
gullibility operator in the sense that it accepts any degree of information upon two operands.
In bilattices, therefore, when the gullibility operator ⊕ is used, getting more information
106
Contents
pushes the overall belief towards true or false with more degree of information except in
case of contradiction. Figure 5.1 (b) (c) (d) shows different bilattice structures that can
Since the introduction of Fuzzy set theory and Fuzzy logic by Zadeh [176, 177], it became
popular as a formalism for representing imprecise or vague linguistic concepts (e.g. such
as ‘hot’, ‘cold’, ‘fast’, etc.). The basic idea is to introduce a fuzzy membership function
fuzzy set. The membership functions are defined to map an element u to a value within
the interval [0, 1] (i.e. µ(u) → [0, 1] ), thereby assigning exact value makes all elements in
the fuzzy set to be ordered and comparable. Due to this aspect, there has been arguing
Chapter 5
that this makes them inadequate for dealing with incomparable uncertain information [42].
There have been some remedies for this aspect. Noting on the footnote comment of Zadeh
saying ‘in a more general setting, the range of the membership function can be taken to be
a suitable partially ordered set P.’ (p. 359 of [176]), Goguen introduced L-fuzzy set that
uses a membership function that maps an element u to a value in a lattice (i.e. µ(u) → L)
[70]. Interval based representation of degree of membership is also introduced with the
name of interval-valued fuzzy sets (IVFSs, for short) by Gehrke et al. [68]. In IVFSs, an
element u is mapped into a subinterval within (i.e. µ(u) → (vl , vu ) ∈ [0, 1]2 , vl ≤ vu ).
Intuitionistic fuzzy set theory (IFSs, for short) introduced by Atanassov [21] additionally
adopts a non membership function ν, with a weaker constraint ν ≤ 1 − µ (note that, in the
constraint µ + ν ≤ 2, we get fuzzy Belnap set. Unlike IVFSs that requires one to address
values to be vl ≤ vu as lower bound and upper bound of imprecision, IFSs allows one to
107
Contents
Figure 5.2: Lattices (a) L∗ Triangle corresponding to Intuitionistic Fuzzy Logic (and also
can be seen as LI Triangle for Interval-Valued Fuzzy Logic), and (b) L□ Square
corresponding to Fuzzy Belnap Logic.
address the positive and the negative side of an imprecise concept separately. Cornelis et
al. [44], however, showed that IVFSs can also be represented in the form of IFSs, in other
words the two are isomorphic and the truth values in IFSs can be represented as intervals
(e.g. by the mapping f (xl , xu ) = (xl , 1 − xu )). As traditional fuzzy set is used for fuzzy
logic as a measure of uncertainty on a proposition, IVFSs, IFSs and fuzzy Belnap set are
adopted for interval-valued fuzzy logic (IVFL, for short), intuitionistic fuzzy logic (IFL,
for short) and the fuzzy Belnap logic (aka. Fuzzified four-valued logic, FOUR). Following
Goguen [70], IVFL and IFL can be defined on ‘triangle species of lattices denoted LI and
L∗ respectively. Fuzzy Belnap logic can be defined on the ‘square’ lattice denoted L□ .
Therefore, IVFSs, IFSs and Fuzzy Belnap logics are kind of L-fuzzy logic. The formal
Definition 23. (L□ Square lattice for fuzzy Belnap logic) . L□ = (L□ , ≤□ ), where
Definition 24. (L∗ Triangle lattice for IFSs) . L∗ = (L∗ , ≤L∗ ), where L∗ = {(x1 , x2 ) ∈
Definition 25. (LI Triangle lattice for IVFSs) . LI = (LI , ≤LI ), where LI =
{(x1 , x2 ) ∈ [0, 1]2 |x1 ≤ x2 } and [x1 , x2 ] ≤LI [y1 , y2 ] iff x1 ≤ y1 and x2 ≤ y2 .
Figure 5.2 shows the corresponding graphical interpretation of IFL and fuzzy Bel-
108
Contents
nap logic. As it was shown in Figure 5.1 (c) and (d), they correspond to continuous
extension of three valued logic and Belnap logic (Figure 5.2 (a) and (b), respectively). In
epistemic sense, the values (0, 0) corresponds to ‘unknown’, (1, 0) corresponds to ‘true’,
considered to have even more information than ‘definite truth’ or ‘definite false’. In IFL and
IVFL, however, this way of ‘contradiction’ state is not allowed because IFL and IVFL do
not allow epistemic points that is considered to have even more information than ‘definite
Defaults (default assumptions) are statements that can be interpreted as ‘normally, typ-
ically, generally true or false’ as a rule. Contrary to defaults, statements that express
Chapter 5
explicit truth or falsity are called definite rules. In practice, the need to make default
assumptions often occurs in cases where the information at hand is uncertain, incomplete
based on known defaults and definite rules. Therefore, in default reasoning, conclusions
Reiter formalized such reasoning aspects as default logic theory using default rules [143].
In the following we give a brief overview on how rules are expressed in logic programming
This section describes Reiter’s formalization of default logic and an example of default
Definition 26. (Default Theory) [143] . Let ∆ = (D, W ) be a default theory, where W
is a set of logical formulae (rules and facts) also known as the definite rules and D is a set
109
Contents
α:β
of default rules of the form γ , where α is known as the precondition, β is known as the
Any default rule dr ∈ D can be also written as ‘γ ← α, not(¬β)’, where not means
the negation by failure to prove. The interpretation of such rule is that, if the precondition
α is known to be true, and if there were no explicit violations of the justification (facts
and rules that entails ¬β) then it is possible to derive the conclusion γ.
scene as depicted in Figure 5.3 with two cameras observing the upper and lower parts of
an escalator respectively. The scene also shows stairs next to the escalator. Consider the
Assume that Cam1 continuously observes that people appear to be using the stairs
set of facts and rules, by default, the rule ¬escalator_working(T1 ) is satisfied because
serve a crowd in front of the escalator and as soon as of Cam2 generates a set of facts
110
Contents
Ginsberg showed the use of bilattice structure to model default reasoning aspect and ex-
tended the structure to generalized default reasoning framework called multivalued default
logic (aka. prioritized default logic) for artificial reasoning [69]. Ginsberg’s bilattice struc-
tures also inherits the behind philosophy of Belnap logic in the sense that they also adopt
the epistemic states ‘unknown’ and ‘contradictory’. To distinguish definite truth and de-
fault truth value, default truth values assumed to have different amount of truth and
different amount of information are also introduced. Figure 5.4 shows Belnap logic that
has no default truth values (i.e., traditional four-valued logic), default logic and multi-
valued default logic respectively. Based on this truth value setup and the four bilattice
operators, each bilattice operator can be seen as a truth functional binary operator on
those values. Table 5.1 shows (a) the truth table of Belnap logic that has no defaults and
(b) default logic that has default true and default false as epistemic states. Based on the
Chapter 5
truth functional binary operators, inference on the bilattice framework is defined in terms
and a truth assignment labeling each sentence k ∈ K with a truth value and a Bilattice B,
then the closure ϕ of a given query sentence q denoted cl(ϕ)(q), is the truth assignment
In other words, the implication of cl(ϕ)(q) is a functional mapping from the ‘enu-
meration of all sentences’ that can entail (denoted by the symbol ‘|=’) q and its contra-
dictory information ¬q to a ‘truth value in bilattice B’. For example, if ϕ labels sentences
{p, q ← p} ∈ K as true; i.e., ϕ(p) = T and ϕ(q ← p) = T , then cl(ϕ) should also label q as
111
Contents
Figure 5.4: (a) Belnap Logic, FOUR (b) Default Logic (c) Multivalued (Prioritized) De-
fault Logic.
Table 5.1: Truth Table of Bilattice Operators on (a) Belnap (b) Default Logic (The un-
shaded Part is exactly compatible with Belnap).
Definition 29. (Default Inference) . Given a query sentence q and given S and S ′ that
are sets of sentences such that S |= q and S ′ |= ¬q, then the default inference is the truth
⊕ ∧ ⊕ ∧
cldi (ϕ)(q) = u∨[ cl(ϕ)(p)] ⊕ ¬ u∨[ cl(ϕ)(p)]. (5.1)
S|=q p∈S S ′ |=¬q p∈S ′
Informally, Equation 5.1 states that for n sets of sentences S that entails q, we first
collect the lowest upper bound (‘∧’) that every sentence in Si can agree on, then take it
112
Contents
if the result contains more truth than u (unknown) along the partial order ≤t . For each
of these truth values of Si , we evaluate the amount of information and choose the most
informative (certain or specific) one among them using greatest lower bound (‘⊕’) along
the partial order ≤k . We do the same process for all Si′ and by understanding the result
from S ′ as contradictory hypothesis, we again collect the most informative one and apply
the negation operation. Finally, both resulting intermediate values for Si and Si′ are joined
along the partial order ≤k again using greatest lower bound (‘⊕’).
This section describes an inference mechanism for default reasoning using subjective logic.
We will discuss how multiple logical values (i.e., default and definite truth values) can
modeling default reasoning based on subjective logic operators by analyzing the default
Chapter 5
reasoning mechanism on bilattice and identifying corresponding and analogous behaviour.
The core idea of default theory in Definition 26 is on the ability to discard a weaker
belief by updating current belief based on a more specific or stronger belief. Among other
approaches, the main strategy of the billatice based default inference shown in Definition 29
was to prioritize possible states of beliefs in an ordered set and to update lower ordered
belief with higher ordered belief. Especially, the ‘join operator’ ⊕ played an important
role for combining competing truth values, that for example, draws an ‘undecidible’ point
when it fuses two conflicting beliefs at the same level of information (i.e., dfn and dtn in
As explained in Section 3.1, subjective logic uses theoretical elements from the
Dempster-Shafer belief theory that the Dempster’s rule (see Definition 7) plays the central
role for fusing different belief mass assignments. Benferhat et al. [26] introduced the use
113
Contents
of Dempster Shafer belief theory for default reasoning. However, Dempster’s rule has been
criticized mainly because highly contradicting beliefs tend to produce counterintuitive re-
sults [98]. Among others, the critique is further formulated and discussed in the form of
examples by Zadeh [178]. Audun Jøsang et al. [98, 94] also introduced such an information
fusion method called ‘consensus’ operator and showed that it is robust even in the highly
contradicting cases.
information can be fused using subjective logic’s ‘consensus’ operator and contrast to
The consensus operator is designed to ‘fairly’ reflect two opinions in a single opinion.
Informally, this means that, each opinion respect the other as much of their ignorance.
be opinions respectively held by agents A and B about the same state x, and let k =
by γ so that γ = iB A A B
x /ix and γ = 1, when both i and i are exactly 0 (i.e., when A and B
Then wxA,B is called the consensus opinion between wxA and wxB , representing an imaginary
agent [A, B]’s opinion about x, as if that agent represented both A and B. By using the
114
Contents
positives and β number of negatives. Especially, using beta distribution, this can be repre-
sented as Beta(p(x)|α, β). Consider two independent observations observations (αA , β A )
and (αB , β B ). To be ‘fair’, each single observation should be reflected with the same
weight, regardless of the total number of observations. Therefore, the fused observation
should be the observation representing (αA + αB , β A + β B ) that represents again a beta
distribution Beta(p(x)|αA + αB , β A + β B ). According to the opinion to beta distribution
mapping scheme shown in Equation 3.9, this can be formulated as follows.
A A
B B
Beta(p(x)|αA + αB , β A + β B )
Beta(p(x)|α , β )
Beta(p(x)|α , β )
⇓
⇓
⇓
⊕
A B A
α α +αB
bA = αA +β bB = αB +β = bA,B = αA +ααB +β (5.2)
A +2
B +2
A +β B +2
A B A
β β +β B
dA = αA +β
dB = αB +β
dA,B = αA +αβB +β
A +2 B +2 A +β B +2
iA = 2 iB = 2 iA,B = 2
αA +β A +2 αB +β B +2 αA +αB +β A +β B +2
Above (5.2), should be the same as the result derived by the Definition 30. For example,
the consensus calculation of bA,B in opinion space is as follows.
Chapter 5
bA iB +bB iA
bA,B = iA +iB −iA iB
,
by replacing bA , iB , bB , iA and iB with the corresponding elements shown in (5.2).
bA iB + bB iA = αA +α αA 2 αB 2 2(αA +αB )
A +2 αB +αB +2 + αB +αB +2 αA +αA +2 = (αA +β A +2)(αB +β B +2)
2(αA +αB +β A +β B +2)
iA + iB − iA iB = αA +α 2
A +2 + αB +αB +2 − (αA +β A +2)(αB +β B +2) = (αA +β A +2)(αB +β B +2)
2 4
+α )
A B
2(α
(( (((
→ bA,B = bA iB +bB iA ( (A(
(αA +β +2)(α
= ( A B A B
B
+β B +2) αA +αB
iA +iB −iA iB
= αA +αB +β A +β B +2
.
+α +β +β (
2(α +2)
(
(( (
((((
(αA +β A +2)(α B
+β B +2)
(5.3)
dA,B and iA,B can be proven similarly as above. Reminding the discussions in Section 3.1.3,
in fact, the ‘consensus’ operator is applying the bayesian theorem itself. The essence
of bayesian theorem is ‘posterior ∝ prior × likelihood’. As the beta distributions are
‘conjugated family’ of distributions, we know that the multiplication can be simplified by
just adding the corresponding index values as follows.
115
Contents
mA
Θ mB
Θ Dempster’s Non-normalised Consensus
x ∈ 2Θ rule Dempster’s rule operator
A,B
Peter 0.98 0.00 0.490 0.0098 wpeter = (0.492, 0.503, 0.005, 1/3), E=0.494
Paul 0.01 0.01 0.015 0.0003 wPA,B
aul = (0.010, 0.985, 0.005, 1/3), E=0.012
A,B
Mary 0.00 0.98 0.490 0.0098 wM ary = (0.492, 0.503, 0.005, 1/3), E=0.494
Θ 0.01 0.01 0.005 0.0001 -
∅ 0.00 0.00 0.000 0.9800 -
Table 5.2: An Example of Applying Dempster’s Rule and Consensus Operator for Uncer-
tain Belief Fusion [94].
mA
Θ mB
Θ Dempster’s Non-normalised Consensus
x∈2 Θ
rule Dempster’s rule operator
A,B
Peter 0.99 0.00 0.000 0.0000 wpeter = (0.495, 0.505, 0.000, 1/3), E=0.495
Paul 0.01 0.01 1.000 0.0001 wPA,B
aul = (0.010, 0.990, 0.000, 1/3), E=0.010
A,B
Mary 0.00 0.99 0.000 0.0000 wM ary = (0.495, 0.505, 0.000, 1/3), E=0.495
Θ 0.00 0.00 0.000 0.0000 -
∅ 0.00 0.00 0.000 0.9999 -
Table 5.3: An Example of Applying Dempster’s Rule and Consensus Operator for Dogmatic
Belief Fusion - Dempster’s Rule Deriving Counter Intuitive Result [94].
Even though the ‘numerical multiplication’ ‘×’ was used on beta distribution level of
subjective logic is different to the consensus operator. While ‘consensus’ concerns about
different opinions about the ‘same’ proposition, ‘logical multiplication (conjunction)’ deals
with different opinions about ‘dfferent’ propositions, and concerns about the chances that
both may correct. For example, given two different binary frame of concern X = {x, ¬x}
and Y = {y, ¬y}, wxA ∧ wyB concerns about the chances being {xy} among X × Y =
Now, let us again consider the Example 1 in Section 3.1.1, that Zadeh [178] used
to criticize Dempster’s rule. The example deals with a murder case with three suspects;
Peter, Paul and Mary. Assume two conflicting testimonies. In Section 3.1.1, we have
reviewed the result (see Table 3.1) of dempster’s rule when the testimonies have some
amount of uncertainty in the testimony. Similarly, Table 5.2 shows the result of applying
consensus operator to the same example. As shown in Table 5.2, both Dempster’s rule
116
Contents
and consensus operator derived similar results. However, as shown in Table 5.3, if we fuse
more highly contradicting beliefs having no uncertainty, Dempster’s rule derives counter
intuitive result saying that Paul is the suspect. Unlike Dempster’s rule, consensus operator
that reflects both opinions fairly. In the case of ‘absolutely conflicting dogmatic opinions’,
namely definite true wA = (1, 0, 0) and definite false wB = (0, 1, 0), the consensus operator
derives wA,B = (0.5, 0.5, 0) which is also a dogmatic opinion. Cognitively speaking, this
(no matter how we have observed) we have infinite positive observations and also infinite
negative observations at the same time (i.e., Beta(α → ∞, β → ∞) = w(0.5, 0.5, 0)).
Chapter 5
Namely, in this view, full contradiction would be linguistically said as ‘a situation that we
know it could occur or not occur at the same absolutely definite rate 1/2 with no doubt’.
Given this, assume that a person should finally determine whether the proposition x would
occur or not. The person may say it seems ‘could happen with probability 0.5 but also could
not happen with probability 0.5’ or more simply ‘half and half ’.
While in logical view, the ‘conflict’ or ‘contradiction’ simply indicates the logically
inconsistent state itself, where definite true and definite false can arise for the same propo-
sition. In this sense, it most likely indicates logical error or wrong set up in the considered
frame of discernment. Namely, in this view, the focus is on the fact that the resulting
opinion had been derived through both the definite true and the definite false. In some
multi-valued logic formalisms such as Belnap, this state is labeled with a symbol such as
‘⊥’ called ‘(full) contradiction’. Especially, in bilattice based logic frameworks, the ‘con-
tradictory point’ is defined to have even more information than ‘definite truth’ and ‘definite
false’ as shown in Table 5.4 and Figure 5.5. This is due to the restriction that it is based
117
Contents
on a set theory (i.e., poset), in which every element should be ordered in a way it has
both the ‘meet’ and the ‘join’ operators. (Note that, the ‘contradictory’ point is ‘the only
point’ that is defined to have even more information than definite true and false as shown
in Section 5.4.1).
In summary, while the former focusses on the interpretation of a given opinion itself,
the latter rather focusses on the process how an opinion had been derived. To address
this discussion, Audun Jøsang [94] proposed an extra parameter called ‘degree of conflict’.
‘An argument that could be used against our consensus operator, is that it does not give any indi-
cation of possible belief conflict. Indeed, by looking at the result only, it does not tell whether the
original beliefs were in harmony or in conflict, and it would have been nice if it did. A possible
way to incorporate the degree of conflict is to add an extra conflict parameter. This
could for example be the belief mass assigned to ∅ in Non-normalised Dempster’s rule, which in
The conflict parameter would only be relevant for combined belief, and not for original beliefs. A
default value c = −1 could for example indicate original belief, because a default value c = 0 could
be misunderstood as indicating that a belief comes from combined harmonious beliefs, even though
it is an original belief.’
The ‘conflict’ parameter above can be used as an indicator to inform that the opinion
is derived by consensus. Following the setup, fusing definite true wxA = (1, 0, 0, 1/2, −1)
and definite false wxB = (0, 1, 0, 1/2, −1) would introduce the ‘conflict’ parameter cA,B =
1·1+0·0 = 1. Therefore, we get, wxA,B = (0.5, 0.5, 0, 1/2, 1). This gives us the clue how the
opinion was derived. Comparing wxZ = (0.5, 0.5, 0, 1/2, −1) and wxA,B = (0.5, 0.5, 0, 1/2, 1),
while the former indicates one independent observation based opinion and the latter indi-
cates that it was fused by two highly contradicting information, the actual distribution of
them are the same. Namely, once an opinion is derived, one practical interpretation could
118
Contents
be to indicate to the application that there is a conflict situation. However, any subsequent
numerical calculation upon the resulting opinion will be the same, no matter how it had
been derived. Therefore, in this dissertation, we will not explicitly use the ‘conflict’ pa-
rameter but rather simply accept the Bayesian view for dealing with ‘(full) contradictory’
state. Then, we will further discuss this aspect again in Discussion 1 in Section 5.9.1.
In this section we discuss the possible mapping of multiple logic values into opinion space.
We start with noting that the implications of ≤t and ≤k in bilattice are similar to the
concept of truth and ignorance in subjective opinion space when visualized in the opinion
triangle. As shown in Figure 5.4 (b) and (c), dtn and df n indicate different levels of
incomplete truth or falsity. The more certain and specific knowledge is obtained, the
higher level of default values result. The degree of certainty or specificity can be considered
Chapter 5
as degree of information, therefore, the levels of default values can be ordered along the
partial order ≤k . Along the information order ≤k , for each pair of dtn and df n there exist
corresponding undecidible states ∗n . As shown in Figure 5.4 (b) and (c), ∗n are assumed
to have more information than their corresponding dtn and df n . Unknown state U is one
of the undecidible states with zero degree of information. Similarly, the full contradictory
state ⊥, that can be reached via definite true (full belief) and definite false (full disbelief),
In the sense of degree of information, however, assigning even higher degree of infor-
mation than definite true or false to the full contradiction point ⊥ is an interesting aspect
to discuss. In bilattice, the full contradictory state ⊥ is considered to have even more infor-
mation than definite true or false. While in subjective logic, it is again a dogmatic opinion
having full degree of information (no ignorance). As discussed in the previous Section 5.6.2,
(i.e., the conflict parameter c = 1) to explicitly indicate that a fused opinion had been
derived via the definite true and the definite false. Therefore, strictly speaking, the bilat-
119
Contents
tice way of representing ‘full contradiction’ in subjective logic is (0.5, 0.5, 0, a, c = 1). To
distinguish this slight difference, we will denote the maximum undecidible point (namely,
full contradictory point) in bilattice as ⊥bl and ⊥sl = (0.5, 0.5, 0, a, 1) for subjective logic
(henceforce, the subscripts bl and sl denote bilattice and subjective logic respectively).
Except the full contradictory point, the rest of undecidible states ∗n can be defined
in the opinion triangle as of bilattice. Additionally, such ∗n should be able to reach via an
operation as of join operator ⊕bl in bilattice on the partial order ≤k . This aspect can be
modeled with subjective logic consensus operator ⊕sl (see Definition 30). By definition,
when dtn and df n (having the same level of ignorance) are fused with the consensus operator
⊕sl , it always yields an opinion in the middle of opinion triangle with less ignorance. The
only exception to this is the case of fusing definite true and definite false that yields an
opinion in the middle between definite true and definite false again with no ignorance.
For example, if we consider a tiny amount of ignorance1 ε and take (1 − ε, 0, ε) = t′ and
(0, 1 − ε, ε) = f ′ as any default true or default false, then fusing t′ and f ′ in terms of
degree of information in subjective logic t′ ⊕sl f ′ = ∗′ will always draw the values with less
ignorance ε′ < ε (see definition of i in the case of k ̸= 0 in Definition 30). Following is the
formal proof of this aspect.
Premise :
Given two opinions wxA = (1 − ε, 0, ε) and wxB = (0, 1 − ε, ε) , where ε ∈ (0, 1)
2(1 − ε) 2(1 − ε) ε ε
wxA ⊕ wxB = ( , , ) satisfies ε > .
2−ε 2−ε 2−ε 2−ε (5.5)
Proof :
ε
ε> ⇒ ε(2 − ε) > ε ⇒ 2 − ε > 1, by the fact that ε ∈ (0, 1) ⇒ (1, 2) > 1
2−ε
ε
∴ε> □
2−ε
This behavior is exactly the same as what bilattice based default structures capture,
thus, in the sense of ordering along ≤k , t′ ≤k ∗′ and f ′ ≤k ∗′ . The only exception to
this is when definite true and false are fused. Namely, t ⊕sl f = ⊥sl (see Definition 30).
This means that only for this single point there is no exact correspondence but rather an
approximate correspondence. Consequently this point in the opinion triangle is denoted
1
Benferhat et al. [26] used similar idea of introducing ε for Dempster Shafer belief theory based default
reasoning.
120
Contents
Chapter 5
Figure 5.5: (a)The Bilattice and Opinion Triangle Space for Belnap FOUR, (b) The Bi-
lattice and Opinion Triangle Space for Default Logic, (c) The Bilattice and
Opinion Triangle Space for Multivalued Default Logic.
as ⊥sl as depicted in Figure 5.5 (a). Figure 5.5 (a) depicts that the correspondence of the
bilattice used for Belnap logic, FOUR and the opinion triangle for example, by mapping
tbl ≃ tsl = (1, 0, 0, a), fbl ≃ fsl = (0, 1, 0, a), ubl ≃ isl = (0, 0, 1, a) and ⊥bl ≃ ⊥sl =
(0.5, 0.5, 0, a, c = 1). As mentioned in Section 5.6.2, however, considering that the conflict
parameter c does not affect to its subsequent calculation, we will not explicitly use the
conflict parameter in this dissertation. For the default values, following the discussion
above and nontheless values in bilattice are elements of a finite set and the opinion triangle
is of continuous domain, we could pick points along the side edges of the opinion triangle.
However, it is important to note that picking the logical values should be done in the sense
of selecting opinion points to put in a set that satisfies ordering like bilattice. Therefore,
121
Contents
there can be many ways of such mapping as long as we pick values such that:
Figure 5.5 (b) shows an example of such mapping of default true and default false to the
opinion triangle. In the same manner, we can extend such mapping to the generalized
multivalued default logic in which each of the defaults can be considered with different
Now, bearing in mind the default inference mechanism defined on bilattice (see Equa-
tion 5.1), we examined subjective logic operators corresponding to ∧bl , ∨bl , ⊗bl and ⊕bl .
Concerning the semantic interpretation of ∧bl and ∨bl representing the standard logical
role of conjunction and disjunction, subjective logic conjunction (·sl ) and disjunction (⊔sl )
operators are examined. For the ⊗bl and ⊕bl operators representing consensus and gulli-
bility , subjective logic consensus ⊕sl and addition +sl operators are examined. Table 5.4
shows the examination results about the correspondence between each of semantically cor-
responding operator pairs. Interestingly, the interpretation of consensus operator ⊗bl does
not match to the consensus ⊕sl in subjective logic. Rather, consensus ⊕sl exhibits char-
acteristics corresponding to the gullibility operator (⊕bl ) in bilattice. The subjective logic
addition operator +sl showed completely different truth table compared to both the ⊕bl and
the ⊗bl . This is because the operator +sl is about adding any beliefs, subtracting any dis-
belief and averaging ignorance from both operands. Thereby, it tends to force rapid change
of belief towards truth direction. The operator ⊗bl seems not to have any corresponding
operator in subjective logic. In Table 5.4, the black points represent that those values are
identical to the ones in bilattice space. Contrary, the red points indicate that the values
are slightly different from the ones in bilattice space. This is due to the difference that the
bilattice operators are defined on a discrete set and the operators in subjective logic are
122
Contents
Chapter 5
Table 5.4: Comparison of Operators used in cl(ϕ) shown in Equation (5.1) (Black - exact
match and Red - slight different match).
defined on continuous space. However, the semantics of the corresponding operators are
Definition 31. (Conjunction ·sl ) [97] . Let ΘX and ΘY be two frames and let x and
Definition 32. (Disjunction ⊔sl ) [97] . Let ΘX and ΘY be two frames and let x and
123
Contents
The truth functional table of logical conjunction (disjunction) in discrete space should
be a function that is closed to its discrete set of truth values. Therefore, considering the
interpretation of conjunction (disjunction), the binary operator ∧bl (∨bl ) should pick the
greatest upper bound (lowest upper bound) element of given two operands. This forces,
by discrete set of values, as is the case in subjective logic, we would expect values in
are interpreted as intersection (union) of both belief values. This aspect is mainly captured
by the definition of bx∧y (bx∨y ). The amount of ignorance is also captured by ix∧y (ix∨y )
so that it can consider both ignorance values of given two belief values. This aspect is
the main source where the differences of the truth table comes from in Table 5.4. Thus,
and disjunction operators model the meaning of the operators under partial ignorance
correctly, but with an additional aspect that is only meaningful in a continuous space.
Similarly to conjunction and disjunction, the operator ⊕bl is also defined to pick a
value among the given discrete set of values. The selection is done in the sense that it
chooses any information that can be accepted upon both operands. In subjective logic, the
consensus operator ⊕sl sees each of operands as one that have observed continuous amount
of positive and negative evidence, thereby, summing up both observations into one opinion.
As discussed in the previous Section 5.6.3, this is similar to interpreting the semantics of
consensus from a bayesian perspective and it will increase the amount of information but
cannot be restricted to be a discrete value. This aspect is captured by the definition of iA,B
x
124
Contents
in Definition 30, except in the case of dogmatic opinions having no ignorance. Therefore,
the meaning of ⊕bl is modeled also in the sense of partial order ≤k in bilattice, i.e., that
the derived value of given two operands should have more information (less ignorance in
subjective logic). Thus, ⊕sl operator in subjective logic models fusing uncertain beliefs in
Based on the consideration about the semantics of operators shown above, we now
defined the truth assignment and closure operation for default reasoning using subjective
logic.
and a truth assignment labeling each sentence k ∈ K with a truth value and Subjective
Opinion Space O, then clsl (ϕ)(q), the closure ϕ of a given query sentence q, is the truth
Chapter 5
assignment function such that :
Definition 35. (Default Inferencesl ) . Given a query sentence q and given S and S ′
that are sets of sentences such that S |= q and S ′ |= ¬q, then the default inference is the
⊕ ∏ ⊕ ∏
clsldi (ϕ)(q) = u⊔[ clsl (ϕ)(p)] ⊕ ¬ u⊔[ clsl (ϕ)(p)] . (5.7)
S|=q p∈S S ′ |=¬q p∈S ′
and Definition 35 corresponds to the Definition 16 - 4. In this dissertation, we will use the
term ‘Opinion Assignment’ to emphasize the use of subjective opinion. In this chapter,
however, we will use the term ‘Truth Assignment’ instead of the term ‘Opinion Assign-
ment’ also to consider the use of other ‘uncertainty representation formalisms’ (such as
125
Contents
Opinion
In this section, as it was done for subjective logic, we extend the discrete bilattice based
multivalued default logic to L-fuzzy logics in continuous space. We then describe some
properties of L-fuzzy logic representations on bilattice structure and review the possibility
on enabling default reasoning. In the preliminaries section, we have considered IVFL, IFL
and fuzzy Belnap logic in terms of truth value order (≤t ). However, as we have examined in
the previous section, the operators (especially, ⊕bl ) on degree of information (≤k ) play an
important role to model default reasoning aspect as to the case of discrete species bilattice
of multivalued default logic. There has been some work on representing IVFSs, IFSs and
Fuzzy Belnap using bilattice. Following is the definitions of ‘triangle’ and ‘square’ in the
context of bilattice. Arieli et al. introduced the following definitions of ‘square’ bilattice
and
When we set L = [0, 1], this captures the Atanassov’s idea on intuitionistic fuzzy sets
but without imposing the restriction µ + ν ≤ 1. Denoting the join and meet operations of
the complete lattice L by ∧L and ∨L , the following four operators are defined for (x1 , x2 ),
(y1 , y2 ) in L2 ,
126
Contents
The ∧L and ∨L can be defined by a t-norm and a t-conorm in the sense that they
generalize intersection and union in lattice space that can be seen as a metric space that
Chapter 5
satisfies triangle inequity. (note that, a t-norm is a function T : [0, 1] × [0, 1] → [0, 1] that
satisfies commutative, monotonicity, associative and one act as identity element. And the
same is for a t-conorm S : [0, 1] × [0, 1] → [0, 1] by replacing the last constraint with zero
identity constraint). Table 5.5 shows some of well known t-norms and t-conorms. Choosing
a pair of them to use for ∧L and ∨L on the lattice L , we can define meet ∧ and join ∨ for
partial order ≤t , and meet ⊗ and join ⊕ for partial order ≤k on the bilattice L2 according
to Equation 5.8 shown above. Therefore, considering the semantics of the Equation 5.1
(see Definition 29), we can directly apply the same inference scheme to bilattice based
fuzzy Belnap logic. Similarly to square bilattice for fuzzy Belnap logic, triangle for IVFSs
Definition 37. (I(L) Triangle bilattice for IVFSs) . Let L = (L, ≤L ) be a complete
127
Contents
Though the definitions of triangle bilattice for IFSs are not explicitly defined in
their work, we c-storean easily introduce the triangle structure for IFSs following the
Definition 37.
Definition 38. (I ∗ (L) Triangle bilattice for IFSs) . Let L = (L, ≤L ) be a complete
As in the case of square bilattice, by setting L = [0, 1], the structure correspond to
LI and L∗ . However, Arieli et al. also showed that I(L) is in fact not a (pre-) bilattice,
since the substructure is not a lattice because the lub (least upper bound, join ∨k ) of any
two elements does not always exist [19]. This corresponds to the interesting aspect that,
the triangles do not allow explicit representation of the epistemic state ‘contradictory’ in
terms of degree of information (note that opinion triangle of subjective logic has the same
aspect). Therefore, for example, the full truth and full falsity in triangle are not comparable
in terms of degree of information. But still (I(L), ≤k ) is a partially ordered set, therefore
the triangle is very much in the same spirit as bilattices. This property is also same in
the case of I ∗ (L). They have also proved that t-norms and t-conorms for the ≤k -order
etc. However, they showed that any t-norms or t-conorms definable in classical fuzzy set
theory have extensions to IVFSs along the partial order ≤t in a compositional manner.
Due to this aspect, it seems that bilattices are not always the key to model the adequate
properties of IVFL and IFL but is quite much adequate for modeling fuzzy Belnap logic.
Therefore, in the context of bilattice, the default inference scheme, Equation 5.8 can not
be set up on IVFL or IFL. While the meet ⊗ and join ⊕ operators can not be defined
on I(L) and I ∗ (L), however, there is an useful mapping between square L□ and triangle
128
Contents
L∗ . In the work [22], Atanasov, the founder of intuitionistic fuzzy logic, further studied on
the relationship between the ‘triangle’ and the ‘square’, and defined following two bijective
transformations F and G from L□ to L∗ , defined for (x1 , x2 ) ∈ [0, 1]2 such that,
(
(0, 0)
if x1 = x2 = 0 x1 − x2 , x2 ) if x1 ≥ x2
( 2 ) G(x1 , x2 ) = 2 2
F (x1 , x2 ) = x1
, xx11+x
x2
if x1 ≥ x2 ( x1 , x − x1 ) if x1 < x2 .
x +x
( 1 2 2
)
2 2 2
(5.10)
x1 x2 x22
x1 +x2 , x1 +x2 if x1 < x2 .
(5.9)
Then later, Cornelis et al. showed that the bijective mapping does not preserve the
order, therefore not lattice isomorphism [44]. However, as for the triangle perspective inter-
pretation of values in square, it is still useful. Therefore, rather than directly model default
reasoning scheme for IFSs and IFSs, we do reasoning on square bilattice for fuzzy Belnap
logic, then transform the derived result using above Equations (4) (5). Definition 27, Def-
Chapter 5
inition 28 and Definition 29 hold on square L2 , and we will again distinguish the inference
on square, by denoting subscript L2 to the closure operation i.e. clL2 . For I(L) and
di (ϕ)(q)
I ∗ (L), we define following projection function relying on the above two possible mappings.
Definition 39. (clIF∗ (L) F-Interpretation) . Given a reasoning result clL2 on square
di (ϕ)(q)
L2 , the F-Interpretation is the function such that clIF∗ (L) = F (clL2 (ϕ)(q)), where the func-
di
Definition 40. (clIG∗ (L) G-Interpretation) . Given a reasoning result clL2 (ϕ)(q) on
di
square L2 , the G-Interpretation is the function such that clIG∗ (L) = G(clL2 (ϕ)(q)), where
di
Reminding that the IVFSs and IFSs are isomorphic [44], in this chapter, we will
show default reasoning on L2 and its F and G interpretations to IFSs. The interpreta-
tions can give us a shedding insight on comparing the reasoning result of the presented
subjective logic based approach with IFL and fuzzy Belnap logic, because the uncertainty
representation using µ and ν in IFL is pretty much similar to the one of subjective logic.
129
Contents
This section deals with illustrative default reasoning examples for visual surveillance sce-
narios. To verify our approach and also to contrast with L-fuzzy logic based approaches,
we will reuse two examples demonstrated by Shet et al. [153, 154] and one scenario in typ-
ical airport scene that is also inspired by Shet et al. [152]. Then we compare the proposed
default inference approach Equation 5.7 to the one of bilattice based default reasoning and
its extension to L2 . The reasoning on L2 will be also interpreted in the view of L∗ (i.e.
I ∗ (L) psuedo-bilattice of L∗ ) with the interpretations clIF∗ (L) and clIG∗ (L) . In this section,
we will not directly concern about IVFL due to the isomorphism between IVFSs and IFSs.
T ≃ (1, 0, 0)sl ≃ (1, 0)L∗ ≃ (1, 0)L2 F ≃ (0, 1, 0)sl ≃ (0, 1)L∗ ≃ (0, 1)L2
DT ≃ (0.5, 0, 0.5)sl ≃ (0.5, 0)L∗ ≃ (0.5, 0)L2 DF ≃ (0, 0, 5, 0.5)sl ≃ (0, 0.5)L∗ ≃ (0, 0.5)L2
U ≃ (0, 0, 1)sl ≃ (0, 0)L∗ ≃ (0, 0)L2 ∗ ≃ (x, x, 1 − 2x)sl ≃ (x, x)L∗ ≃ (x, x)L2
i.e. undecided or contradiction with some amount of uncertainty (degree of information) on opinion
triangle (on L-fuzzy sets), and ⊥ ≃ (0.5, 0.5, 0)sl ≃ (0.5, 0.5)L∗ ≃ (1, 1)L2 (i.e. the full contradic-
tion).
These mappings are reasonable in the sense that the uncertainty representation of opinion
triangle and IFL I ∗ (L) are similar except the atomicity value of opinion triangle. For the
simplicity we assume that all the propositional knowledge we consider are balanced, there-
fore we set the atomicity of opinion triangle as default a = 0.5, and we will not explicitly
denote a. For the rest of truth values we will use opinion triple representation (b, d, i).
For values of opinion triangle, I ∗ (L) and L2 , that are slightly different to above symbols
but still can be interpreted as one of them, we will denote it with superscript ′ . (e.g.
∗ ≃ ∗′ , DT ≃ DT ′ , etc.).
Example 3. (Identity Inference) [153, 154] . Assume the following truth assignment
and set of rules about determining whether two individuals observed in an image should
130
Contents
ϕ[¬equal(P1 , P2 ) ← distinct(P1 , P2 )] = DT
ϕ[equal(P1 , P2 ) ← appear_similar(P1 , P2 )] = DT
ϕ[appear_similar(a, b)] = T
ϕ[distinct(a, b)] = T
Given two default true rules and facts that can be seen as definite true, the inference for
default logic shown in [153] with bilattice and the default inference with subjective logic
are as follows.
= [U ∨ (T ∧ DT )] ⊕ ¬[U ∨ (T ∧ DT )] = [U ∨ DT ] ⊕ ¬[U ∨ DT ] = DT ⊕ DF = ∗
Chapter 5
= (0.67, 0, 0.33) ⊕ ¬(0.67, 0, 0.33) = (0.67, 0, 0.33) ⊕ (0, 0.67, 0.33) = (0.4, 0.4, 0.2) = ∗′
And as shown in Table 5.5, choosing one of t-norm and t-conorm pair, and applying
Equation 5.8 in Definition 36, we get following inference result derived on L2 , and its
interpretations clIF∗ (L) and clIG∗ (L) on I ∗ (L). The reasoning results are as follows.
min / max
clL2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ (min(1, 0.5), max(0, 0))] ⊕ ¬[(0, 0) ∨ (min(1, 0.5), max(0, 0))]
= [(0, 0) ∨ (0.5, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)] = [max(0, 0.5), min(0, 0)] ⊕ ¬[max(0, 0.5), min(0, 0)]
= (0.5, 0) ⊕ ¬(0.5, 0) = (0.5, 0) ⊕ (0, 0.5) = (max(0.5, 0), max(0, 0.5)) = (0.5, 0.5) = ∗′
2
) = (0.25, 0.25) = ∗′
min / max 0.5 0.5·0.5
clIF∗ (L) (clL2 (ϕ)(equal(a, b))) = clIF∗ (L) ((0.5, 0.5)) = ( 0.5+0.5 , 0.5+0.5
di
min / max
clIG∗ (L) (clL2 (ϕ)(equal(a, b))) = clIG∗ (L) ((0.5, 0.5)) = (0.5 − 0.5 0.5
2 , 2 ) = (0.25, 0.25) = ∗
di
prod/sum
clL2 (ϕ)(equal(a, b))
di
131
Contents
= [(0, 0) ∨ (1 · 0.5, 0 + 0 − 0 · 0)] ⊕ ¬[(1 · 0.5, 0 + 0 − 0 · 0)] = [(0, 0) ∨ (0.5, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)]
= [0 + 0.5 − 0 · 0.5, 0 · 0] ⊕ ¬[0 + 0.5 − 0 · 0.5, 0 · 0] = (0.5, 0) ⊕ ¬(0.5, 0) = (0.5, 0) ⊕ (0, 0.5)
Luk
clL2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ (max(0, 1 + 0.5 − 1), min(0 + 0, 1))] ⊕ ¬[(0, 0) ∨ (max(0, 1 + 0.5 − 1), min(0 + 0, 1))]
= [(0, 0) ∨ (0.5, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)] = [min(0 + 0.5, 1), max(0, 0 + 0 − 1)]
⊕ ¬[min(0 + 0.5, 1), max(0, 0 + 0 − 1)] = (0.5, 0) ⊕ ¬(0.5, 0) = (0.5, 0) ⊕ (0, 0.5)
drastic
clL2 (ϕ)(equal(a, b))
di
= ([U ∨ (T ∧ DT )] ⊕ ¬[U ∨ (T ∧ DT )] = [(0, 0) ∨ {(1, 0) ∧ (0.5, 0)}] ⊕ ¬[(0, 0) ∨ {(1, 0) ∧ (0.5, 0)}]
= [(0, 0) ∨ (TD (1, 0.5), SD (0, 0))] ⊕ ¬[(0, 0) ∨ (TD (1, 0.5), SD (0, 0))] = [(0, 0) ∨ (0.5, 0)]
⊕ ¬[(0, 0) ∨ (0.5, 0)] = [SD (0, 0.5), TD (0, 0)] ⊕ ¬[SD (0, 0.5), TD (0, 0)]
= (0.5, 0) ⊕ ¬(0.5, 0) = (0.5, 0) ⊕ (0, 0.5) = (SD (0.5, 0), SD (0, 0.5)) = (0.5, 0.5) = ∗′
′
clIF∗ (L) (clL 2 (ϕ)(equal(a, b))) = clI ∗ (L) ((0.5, 0.5)) = (0.25, 0.25) = ∗
Luk F
di
′
clIG∗ (L) (clL 2 (ϕ)(equal(a, b))) = clI ∗ (L) ((0.5, 0.5)) = (0.25, 0.25) = ∗
Luk G
di
N ilpotent
clL2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ (TnM (1, 0.5), SnM (0, 0))] ⊕ ¬[(0, 0) ∨ (TnM (1, 0.5), SnM (0, 0))]
= [(0, 0) ∨ (0.5, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)] = [SnM (0, 0.5), TnM (0, 0)] ⊕ ¬[SnM (0, 0.5), TnM (0, 0)]
= (0.5, 0) ⊕ ¬(0.5, 0) = (0.5, 0) ⊕ (0, 0.5) = (SnM (0.5, 0), SnM (0, 0.5)) = (0.5, 0.5) = ∗′
N ilpotent
clIF∗ (L) (clL 2 (ϕ)(equal(a, b))) = clIF∗ (L) ((0.5, 0.5)) = (0.25, 0.25) = ∗′
di
N ilpotent
clIG∗ (L) (clL 2 (ϕ)(equal(a, b))) = clIG∗ (L) ((0.5, 0.5)) = (0.25, 0.25) = ∗′
di
132
Contents
Figure 5.6: Reasoning Results of Example 3 in Opinion Space and L2 & I ∗ (L).
Chapter 5
Hamacher
clL 2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ (TH0 (1, 0.5), SH2 (0, 0))] ⊕ ¬[(0, 0) ∨ (TH0 (1, 0.5), SH2 (0, 0))]
= [(0, 0) ∨ (0.5, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)] = [SH2 (0, 0.5), TH0 (0, 0)] ⊕ ¬[SH2 (0, 0.5), TH0 (0, 0)]
= (0.5, 0) ⊕ ¬(0.5, 0) = (0.5, 0) ⊕ (0, 0.5) = (SH2 (0.5, 0), SH2 (0, 0.5)) = (0.5, 0.5) = ∗′
Figure 5.6 shows the graphical representation of above reasoning results. The resulting
opinion (0.4, 0.4, 0.2) represents same amount of degree of truth and false with uncertainty.
This can also be represented as undecided state ∗′ . All reasoning results on L2 also yielded
similar result (0.5, 0.5) and the same F/G interpretations (0.25, 0.25). Thus, the semantics
of results from the discrete bilattice for multivalued default logic, bilattice based L-fuzzy
logics and subjective logic are the same. While the uncertainty representation semantics
of subjective logic is similar to IFL, when the reasoning result on fuzzy Belnap logic is
133
Contents
interpreted, the distance between subjective opinion and the value of IFL was bigger than
Example 4. (Identity Inference with contextual cues) [153, 154] . Assume that
a person enters an office room that we believe to be empty and closed (no other exit).
Suppose that after a while, another person appears from the room who seems dissimilar
from the first person. In this case, inferring equality based on appearance matching is a
weaker default than inferring equality based on the fact that person entered and exited an
empty closed world. This aspect can be represented as following truth assignment and set
of rules.
ϕ[equal(P1 , P2 ) ← enter_closed_world(P1 , X, T1 ),
exit_closed_world(P2 , X, T2 ), T2 > T1 ,
ϕ[¬appear_similar(a, b)] = T
The Inference in this setup is multivalued default reasoning and the bilattice based infer-
ence result shown in Shet et al. [153] and the result of subjective logic based inference are
as follows.
= [U ∨(T ∧T ∧T ∧T ∧T ∧DT2 )]⊕¬[U ∨(T ∧DT1 )] = [U ∧DT2 ]⊕¬[U ∧DT1 ] = DT2 ⊕DF1 = DT2
134
Contents
= (0.9, 0, 0.1) ⊕ ¬(0.67, 0, 0.33) = (0.9, 0, 0.1) ⊕ (0, 0.67, 0.33) = (0.75, 0.17, 0.08) = DT2′
Choosing a pair of t-norm and t-conorm, inference result derived on L2 , and its interpre-
min / max
clL2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ ((1, 0) ∧ (0.8, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] = [(0, 0) ∨ (min(1, 0.8), max(0, 0))]
⊕ ¬[(0, 0) ∨ (min(1, 0.5), max(0, 0))] = [(0, 0) ∨ (0.8, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)]
= (max(0, 0.8), min(0, 0)) ⊕ ¬(max(0, 0.5), min(0, 0)) = (0.8, 0) ⊕ ¬(0.5, 0) = (0.8, 0) ⊕ (0, 0.5)
Chapter 5
prod/sum
clL2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ ((1, 0) ∧ (0.8, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] = [(0, 0) ∨ (1 · 0.8, 0 + 0 − 0 · 0)]
= (0 + 0.8 − 0 · 0.8, 0 · 0)) ⊕ ¬(0 + 0.5 − 0 · 0.5, 0 · 0) = (0.8, 0) ⊕ ¬(0.5, 0) = (0.8, 0) ⊕ (0, 0.5)
Luk
clL 2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ (max(0, 1 + 0.8 − 1), min(0 + 0, 1))] ⊕ ¬[(0, 0) ∨ (max(0, 1 + 0.5 − 1), min(0 + 0, 1))]
= [(0, 0) ∨ (0.8, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)] = (min(0 + 0.8, 1), max(0, 0 + 0))
⊕ ¬(min(0 + 0.5, 1), max(0, 0 + 0)) = (0.8, 0) ⊕ ¬(0.5, 0) = (0.8, 0) ⊕ (0, 0.5)
135
Contents
drastic
clL2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ ((1, 0) ∧ (0.8, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] = [(0, 0) ∨ (TD (1, 0.8), SD (0, 0))]
⊕ ¬[(0, 0) ∨ (TD (1, 0.5), SD (0, 0))] = [(0, 0) ∨ (0.8, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)]
= (SD (0, 0.8), TD (0, 0)) ⊕ ¬(SD (0, 0.5), TD (0, 0))(0.8, 0) ⊕ ¬(0.5, 0) = (0.8, 0) ⊕ (0, 0.5)
N ilpotent
clL2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ ((1, 0) ∧ (0.8, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] = [(0, 0) ∨ (TnM (1, 0.8), SnM (0, 0))]
⊕ ¬[(0, 0) ∨ (TnM (1, 0.5), SnM (0, 0))] = [(0, 0) ∨ (0.8, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)]
= (SnM (0, 0.8), TnM (0, 0)) ⊕ ¬(SnM (0, 0.5), TnM (0, 0)) = (0.8, 0) ⊕ ¬(0.5, 0) = (0.8, 0) ⊕ (0, 0.5)
N ilpotent
clIG∗ (L) (clL 2 (ϕ)(equal(a, b))) = clIG∗ (L) ((0.8, 0.5)) = (0.55, 0.25) = DT2′
di
Hamacher
clL2 (ϕ)(equal(a, b))
di
= [(0, 0) ∨ ((1, 0) ∧ (0.8, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] = [(0, 0) ∨ (TH0 (1, 0.8), SH2 (0, 0))]
⊕ ¬[(0, 0) ∨ (TH0 (1, 0.5), SH2 (0, 0))] = [(0, 0) ∨ (0.8, 0)] ⊕ ¬[(0, 0) ∨ (0.5, 0)]
= (SH2 (0, 0.8), TH0 (0, 0)) ⊕ ¬(SH2 (0, 0.5), TH0 (0, 0)) = (0.8, 0) ⊕ ¬(0.5, 0)
= (0.8, 0) ⊕ (0, 0.5) = (SH2 (0.8, 0), SnM (0, 0.5)) = (0.8, 0.5) = DT2′
136
Contents
Figure 5.7: Reasoning Results of Example 4 in Opinion Space and L2 & I ∗ (L).
Chapter 5
Figure 5.7 shows above reasoning results. The opinion labeled (a) in opinion triangle is
the most closest one to DT2 compared with L-fuzzy logic based results but a bit biased
to center than original DT2 . In the sense of truth value, this is the same to the cases of
yielded values on L2 . However, in these cases, the amount of falsity is also relatively high
therefore the point is located in the area of overflowed information (µ + ν > 1). The F and
G interpretations are also a bit different but very close to each other. Nonetheless, in the
sense that all reasoning results are pointed on the right-hand side of the line of undecidable
*, semantically, this can be interpreted as the meaning of week truth like DT2 . Thus, the
semantics of results from the discrete bilattice for multivalued default logic, the bilattice
137
Contents
Example 5. (Theft Inference with contextual cues) (The scenario and rules for
‘theft’ have been inspired by Shet et al. [152]) . Assume a typical airport surveillance as
depicted in Figure 5.8 with two cameras. Suppose that a human P1 carrying an object
(Baggage) B is observed and stayed around telephone booth in Cam1. After a while he
disappears from the view of Cam1 without taking his baggage B. Subsequently, P2 enters
the scene, picks up the baggage and leaves. In parallel, according to Cam2 , it seems that
P1 and P2 belong to a same group of people so the two people are considered as friends.
In this scenario, based on the possession relation between an object and person, we could
build a default rule to infer whether a person is a thief or not. Similarly, based on the
friend relation we can also build a bit stronger default rule saying possessing object of
friend is not thief. This aspect is depicted as following truth assignment and set of rules.
ϕ[human(P )] = T
ϕ[package(B)] = T
ϕ[possess(P, B, T )] = T
ϕ[¬(belongs(B, P, T ))] = T
ϕ[belongs(B, P ′ , T )]cam2 = T
Given above rules and facts (gathered till P2 is picking up the baggage B), inferring
138
Contents
Choosing a pair of t-norm and t-conorm, inference result derived on L2 , and its interpre-
Chapter 5
tations clIF∗ (L) and clIG∗ (L) on I ∗ (L) are as follows.
min / max
clL2 (ϕ)(thef t(P, B, T ))
di
= [(0, 0) ∨ (min(1, 0.5), max(0, 0))] = [(0, 0) ∨ (0.5, 0)] = (max(0, 0.5), min(0, 0))
= (0.5, 0) = DT1
min / max 2
clIF∗ (L) (clL2 (ϕ)(thef t(P, B, T ))) = clIF∗ (L) ((0.5, 0)) = ( 0.5 0.5·0
0.5 , 0.5+0 ) = (0.5, 0) = DT1
di
min / max
clIG∗ (L) (clL2 (ϕ)(thef t(P, B, T ))) = clIG∗ (L) ((0.5, 0)) = (0.5 − 02 , 02 ) = (0.5, 0) = DT1
di
prod/sum
clL2 (ϕ)(thef t(P, B, T ))
di
= [(0, 0) ∨ (1 · 0.5, 0 + 0 − 0 · 0)] = [(0, 0) ∨ (0.5, 0)] = (0 + 0.5 − 0 · 0.5), 0 · 0) = (0.5, 0) = DT1
prod/sum
clIF∗ (L) (clL2 (ϕ)(thef t(P, B, T ))) = clIF∗ (L) ((0.5, 0)) = (0.5, 0) = DT1
di
prod/sum
clIG∗ (L) (clL2 (ϕ)(thef t(P, B, T ))) = clIG∗ (L) ((0.5, 0)) = (0.5, 0) = DT1
di
Luk
clL 2 (ϕ)(thef t(P, B, T ))
di
139
Contents
drastic
clL2 (ϕ)(thef t(P, B, T ))
di
= [(0, 0) ∨ (TD (1, 0.5), SD (0.5, 0))] = [(0, 0) ∨ (0.5, 0)] = (SD (0, 0.5), TD (0, 0)) = (0.5, 0) = DT1
N ilpotent
clL2 (ϕ)(thef t(P, B, T ))
di
= [(0, 0) ∨ (TnM (1, 0.5), SnM (0.5, 0))] = [(0, 0) ∨ (0.5, 0)] = (SnM (0, 0.5), TnM (0, 0))
= (0.5, 0) = DT1
N ilpotent
clIF∗ (L) (clL2 (ϕ)(thef t(P, B, T ))) = clIF∗ (L) ((0.5, 0)) = (0.5, 0) = DT1
di
N ilpotent
clIG∗ (L) (clL2 (ϕ)(thef t(P, B, T ))) = clIG∗ (L) ((0.5, 0)) = (0.5, 0) = DT1
di
Hamacher
clL2 (ϕ)(thef t(P, B, T ))
di
= [(0, 0) ∨ (TH0 (1, 0.5), SH2 (0.5, 0))] = [(0, 0) ∨ (0.5, 0)] = (SH2 (0, 0.5), TH0 (0, 0))
= (0.5, 0) = DT1
Relying only the facts generated by Cam1 (those are not explicitly subscripted with cam-
era id), all of above inference concluded cl(ϕ)(thef t(P2 , B, T )) = DT1 except subjective
′
logic based approach that derived DT2 . However, the semantic interpretation of the results
are all the same. Namely. theft has taken place with low confidence.
140
Contents
Choosing a pair of t-norm and t-conorm, inference result derived on L2 , and its interpre-
min / max
clL2 (ϕ)(thef t(P, B, T ))
di
= [U ∨ (T ∧ DT1 )] ⊕ ¬[U ∨ (T ∧ DT2 )] = [(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.8, 0))]
Chapter 5
= [(0, 0) ∨ (min(1, 0.5), max(0, 0))] ⊕ ¬[(0, 0) ∨ (min(1, 0.8), max(0, 0))]
= [(0, 0) ∨ (0.5, 0)] ⊕ ¬[(0, 0) ∨ (0.8, 0)] = (max(0, 0.5), min(0, 0)) ⊕ ¬(max(0, 0.8), min(0, 0))
= (0.5, 0) ⊕ ¬(0.8, 0) = (0.5, 0) ⊕ (0, 0.8) = (max(0.5, 0), max(0, 0.8)) = (0.5, 0.8) = DF2′
2
) = (0.3, 0.49) = DF1′
min / max 0.5·0.8 0.8
clIF∗ (L) (clL2 (ϕ)(thef t(P, B, T ))) = clIF∗ (L) ((0.5, 0.8)) = ( 0.5+0.8 , 0.5+0.8
di
prod/sum
clL2 (ϕ)(thef t(P, B, T ))
di
= [U ∨ (T ∧ DT1 )] ⊕ ¬[U ∨ (T ∧ DT2 )] = [(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.8, 0))]
= [(0, 0)∨(1·0.5, 0+0−0·0)]⊕¬[(0, 0)∨(1·0.8, 0+0−0·0)] = [(0, 0)∨(0.5, 0)]⊕¬[(0, 0)∨(0.8, 0)]
= (0 + 0.5 − 0 · 0.5, 0 · 0) ⊕ ¬(0 + 0.8 − 0 · 0.8, 0 · 0) = (0.5, 0) ⊕ ¬(0.8, 0) = (0.5, 0) ⊕ (0, 0.8)
(ϕ)(thef t(P, B, T ))) = clIF∗ (L) ((0.5, 0.8)) = (0.3, 0.49) = DT2′
prod/sum
clIF∗ (L) (clL2
di
(ϕ)(thef t(P, B, T ))) = clIG∗ (L) ((0.5, 0.8)) = (0.25, 0.55) = DT2′
prod/sum
clIG∗ (L) (clL2
di
141
Contents
Luk
clL2 (ϕ)(thef t(P, B, T ))
di
= [U ∨ (T ∧ DT1 )] ⊕ ¬[U ∨ (T ∧ DT2 )] = [(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.8, 0))]
= [(0, 0) ∨ (max(0, 1 + 0.5 − 1), min(0 + 0, 1)] ⊕ ¬[(0, 0) ∨ (max(0, 1 + 0.8 − 1), min(0 + 0, 1)]
= (min(0 + 0.5, 1), max(0, 0 + 0 − 1)) ⊕ ¬(min(0 + 0.8, 1), max(0, 0 + 0 − 1))
= (0.5, 0) ⊕ ¬(0.8, 0) = (0.5, 0) ⊕ (0, 0.8) = (min(0 + 0.5, 1), min(0 + 0.8, 1)) = (0.5, 0.8) = DF2′
′
clIF∗ (L) (clL
Luk F
2 (ϕ)(thef t(P, B, T ))) = clI ∗ (L) ((0.5, 0.8)) = (0.3, 0.49) = DT2
di
drastic
clL2 (ϕ)(thef t(P, B, T ))
di
= [U ∨ (T ∧ DT1 )] ⊕ ¬[U ∨ (T ∧ DT2 )] = [(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.8, 0))]
= [(0, 0) ∨ (TD (1, 0.5), SD (0, 0)] ⊕ ¬[(0, 0) ∨ (TD (1, 0.8), SD (0, 0)]
= [(0, 0) ∨ (0.5, 0)] ⊕ ¬[(0, 0) ∨ (0.8, 0)] = (SD (0, 0.5), TD (0, 0)) ⊕ ¬(SD (0, 0.8), TD (0, 0))
= (0.5, 0) ⊕ ¬(0.8, 0) = (0.5, 0) ⊕ (0, 0.8) = (SD (0, 0.5), SD (0, 0.8)) = (0.5, 0.8) = DF2′
N ilpotent
clL2 (ϕ)(thef t(P, B, T ))
di
= [U ∨ (T ∧ DT1 )] ⊕ ¬[U ∨ (T ∧ DT2 )] = [(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.8, 0))]
= [(0, 0) ∨ (TnM (1, 0.5), SnM (0, 0)] ⊕ ¬[(0, 0) ∨ (TnM (1, 0.8), SnM (0, 0)]
= [(0, 0) ∨ (0.5, 0)] ⊕ ¬[(0, 0) ∨ (0.8, 0)] = (SnM (0, 0.5), TnM (0, 0)) ⊕ ¬(SnM (0, 0.8), TnM (0, 0))
= (0.5, 0) ⊕ ¬(0.8, 0) = (0.5, 0) ⊕ (0, 0.8) = (SnM (0, 0.5), SnM (0, 0.8)) = (0.5, 0.8) = DF2′
N ilpotent
clIF∗ (L) (clL 2 (ϕ)(thef t(P, B, T ))) = clIF∗ (L) ((0.5, 0.8)) = (0.3, 0.49) = DT2′
di
N ilpotent
clIG∗ (L) (clL 2 (ϕ)(thef t(P, B, T ))) = clIG∗ (L) ((0.5, 0.8)) = (0.25, 0.55) = DT2′
di
Hamacher
clL2 (ϕ)(thef t(P, B, T ))
di
142
Contents
Figure 5.9: Reasoning Results of Example 5 in Opinion Space and L2 & I ∗ (L).
Chapter 5
= [U ∨ (T ∧ DT1 )] ⊕ ¬[U ∨ (T ∧ DT2 )] = [(0, 0) ∨ ((1, 0) ∧ (0.5, 0))] ⊕ ¬[(0, 0) ∨ ((1, 0) ∧ (0.8, 0))]
= [(0, 0) ∨ (TH0 (1, 0.5), SH2 (0, 0)] ⊕ ¬[(0, 0) ∨ (TH0 (1, 0.8), SH2 (0, 0)] = [(0, 0) ∨ (0.5, 0)]
⊕ ¬[(0, 0) ∨ (0.8, 0)] = (SH2 (0, 0.5), TH0 (0, 0)) ⊕ ¬(SH2 (0, 0.8), TH0 (0, 0)) = (0.5, 0) ⊕ ¬(0.8, 0)
= (0.5, 0) ⊕ (0, 0.8) = (SH2 (0, 0.5), SH2 (0, 0.8)) = (0.5, 0.8) = DF2′
Figure 5.9 shows above reasoning results. In case of Case1, subjective logic yielded a
rather strong opinion (0.74, 0, 0.26) that is rather close to DT2 than DT1 that other ap-
proaches yielded. Contrary to Case 1, when we take more information also from Cam2,
all of above inference concluded cl(ϕ)(thef t(P2 , B, T )) = DF2 and DF2′ . Namely, no theft
has taken place with rather high confidence. The opinion DF2′ (0.22, 0.7, 0.08) in opinion
triangle is the most closest one to DF2 compared to L-fuzzy logic based results. In the
sense of truth value, this is the same to the cases of yielded values on L2 . However, in
this case, the amount of falsity is also relatively high therefore the point is located in the
143
Contents
area of overflowed information (µ + ν > 1). The F and G interpretations are both close to
each other however the interpretation is rather closer to DF1 . Nonetheless, in the sense
that all reasoning results are pointed on the left-hand side of the line of undecidable *,
semantically, this can be interpreted as the meaning of week false like DF2 . Thus, the
semantics of results from the discrete bilattice for multivalued default logic, the bilattice
As shown with above illustrative visual surveillance inference scenarios, the proposed
default reasoning mechanism Equation 5.7 semantically well models default reasoning and
5.9 Discussion
As shown in the previous section with examples of default reasoning in visual surveillance,
both subjective logic and L-fuzzy logics (especially, IFL and fuzzy Belnap logic) seem rel-
evant for the use of approximate default reasoning. What makes IFSs attractive compared
to other fuzzy set extensions is that it makes geometrical interpretations possible, thereby,
Euclidean plane with a Cartesian coordinate system. [23]. This aspect is also the same
in Subjective logic because subjective logic also makes the geometrical interpretations of
Beta probability distributions possible in the Euclidean plane. However, there are still
some properties worth to discuss to contrast these approaches. In this section, we give
a comparison on the property of both approaches in the view of logical soundness and
As we noted in the previous default reasoning section, the initial idea of discrete bilattice
assumed an epistemic state called ‘unknown’ and ‘contradiction’ (defined to have even more
information than definite true or false) following Belnaps four-valued logic. In the sense
144
Contents
of discrete valued logic such as multivalued default logic, the only way of reaching to the
overflowed information state is through the handling of definite true and definite false, and
‘contradiction’ ⊥ is the only epistemic state that is defined in the overflowed information
area. However, as shown in the previous examples, this is not true on L2 , because we can
encounter values in the area of (µ + ν > 1). Similarly to the discussion we have reviewed in
Section 5.6.2, regarding the meaning of the area, and more specifically the epistemic state
‘contradiction’, there has been many discussions on the significance of the epistemic state
in the view of logic. A critical review can be found in a work of Urquhart [166]. There has
been also some report on the possible problems that could be introduced for the formal
Dubois, D. [56] formally showed the problems that can arise in Belnap logic, in the sense
of logical soundness. In the following, we briefly review the discussions introduced by [56].
Chapter 5
(a) and (b) show the truth table of the Belnap’s four-valued logic and the default logic.
In both logics, the conjunction and disjunction are problematic when applied to the two
extreme epistemic states ‘Unknown’ U and ‘Contradiction’ ⊥. For instance in Belnap logic,
consider p ∧ ¬p and p ∨ ¬p. In Belnap logic, the former and latter both are U because
former should be false and the latter should be true. The same anomaly is also introduced
p ∧ ¬q and p ∨ ¬q, again, we get ⊥ for both. It breaks the tautology in classical logic sense
From a common sense, the results are counterintuitive and this was even to Belnap
because he stated that this is an unavoidable consequence of his formal setting [56]. This
aspect can be problematic in the following scenario. Consider a proposition p and q, and
agent A1 and A2 saying the p is T and F respectively and this is why p is ⊥. Because A1
145
Contents
counter intuitive [56]. This aspect also leads to debatable epistemic value assignments.
Suppose two atomic propositions p and q with epistemic state assignment ϕ(p) = ⊥ and
ϕ(q) = U . Then ϕ(p ∧ q) = F as noted above. But since Belnap negation is such that
ϕ(¬p) = ⊥ and ϕ(¬q) = U , we also get ϕ(¬p ∧ q) = ϕ(p ∧ ¬q) = ϕ(¬p ∧ ¬q) = F . Hence,
ϕ((p∧q)∧(¬p∧q)∧(p∧¬q)∧(¬p∧¬q)) = ϕ(p∧q)∧ϕ(¬p∧q)∧ϕ(p∧¬q)∧ϕ(¬p∧¬q) = F ,
This aspect shows that, for any logical connectives ∗, ϕ(p) ∗ ϕ(q) ̸= ϕ(p ∗ q) in Belnap
each proposition can not characterize a single epistemic value for the combination of the
propositions. This aspect also hold in the fuzzy Belnap logic as well, because regardless
what t-norms and t-conorms we choose, the truth table values corresponding to definite
true, definite false, unknown and contradictory values will have the same truth functional
Unlike fuzzy Belnap logic, in subjective logic we can avoid this problem by the use
of atomicity value a, therefore subjective logic better captures the spirit of classical logic2 .
Consider the same case of U ∧ U and U ∨ U . As shown in Figure 5.10 (a), for subjective
logic opinion w = (0, 0, 1, 0.5) which corresponds to U , subjective logic conjunction also
draws full ignorance but with different atomicity, namely (0, 0, 1, 0.5) ∧ ¬(0, 0, 1, 0.5) =
(0, 0, 1, 0.25). The semantics is clear. Namely, for a proposition that is known to be binary
event that an agent has a full ignorance on the truth of it, the conjunction also draws
full uncertainty but, following the spirit of probabilistic conjunction, it comes with the
atomicity that is the product of both atomicity (i.e. 0.5 · 0.5 in this case). Therefore, even
if we get full ignorance, when it is interpreted in terms of Beta distribution, the overall
expectation should be biased to falsity as traditional logic yields F . This is the same in
the case of U ∧ U that yields a full ignorance opinion but its atomicity is biased to T . The
2
Strictly speaking, subjective logic does not allow a ‘truly full uncertain’ state, in the sense that subjective
logic adds additional dimension of information called atomicity (prior) value. This is the main cause
of the difference shown in Figure 5.10 (a).
146
Contents
Chapter 5
Figure 5.10: (a) U ∧ U and U ∨ U (b) ⊥ ∧ ⊥ and ⊥ ∨ ⊥ in Subjective Logic.
classical logic should draw T and F . As shown in Figure 5.10 (b), considering the epistemic
state of contradiction as w = (0.5, 0.5, 0, 0.5), we get (0.5, 0.5, 0, 0.5) ∧ ¬(0.5, 0.5, 0, 0.5) =
(0.25, 0.75, 0, 0.25) that is biased to disbelief. Note that, both in Figure 5.10 (a) and (b)
we have the same probability expectation values. However, when represented as Beta
3
This is due to the main difference that subjective logic offers the Bayesian view on interpreting contra-
diction as discussed in Section 5.6.2. Strictly speaking, Belnap like logical view of contradiction can be
also explicitly distinguished using the ‘conflict parameter’ in subjective logic (see Section 5.6.2), even
though the use of the parameter does not affect to subsequent calculation and therefore we did not
explicitly use the parameter in this dissertation. However, for some applications in which it is very
important to know and to distinguish the occurrence of the logical view of conflict, so to regard it
as unhealthy system status and therefore, for example to halt the system, we could explicitly use the
‘conflict’ parameter.
147
Contents
Figure 5.11: (a) Two Examples of U ∧ ⊥ (b) Two Examples of U ∨ ⊥ in Subjective Logic.
distribution, while (b) is almost certain because we have rather pick distribution (b) is
almost uncertain. This aspect is directly captured in the opinion triangle by the value of
logic, subjective logic draws a bit different epistemic states. Figure 5.11 (a) depicts two
cases of U ∧ ⊥ one more biased to F and the other more biased to ⊥. The basic idea is
that we take more atomicity in the case of unknown opinion. The same aspect is captured
in the case of U ∨ ⊥ as shown in Figure 5.11 (b). Figure 5.11 (a) more intuitively explains
the above mentioned agent scenario with two propositions p and q. Again, consider the
For the proposition q, we assign U because agent A1 and A2 say nothing about q. Now,
148
Contents
in Belnap logic, has no problem in subjective logic [98, 97, 94]. Suppose two atomic
propositions p and q with epistemic state assignment ϕ(p) = ⊥ = (0.5, 0.5, 0, 0.5) and
ϕ(q) = U = (0, 0, 1, 0.5). Then, ϕ(p ∧ q) = (0.17, 0.5, 0.33, 0.25). The negation in
this case is ϕ(¬p) = ⊥ = (0.5, 0.5, 0, 0.5) and ϕ(¬q) = U = (0, 0, 1, 0.5), we also get
ϕ(¬p ∧ q) = ϕ(p ∧ ¬q) = ϕ(¬ ∧ ¬q) = (0.17, 0.5, 0.33, 0.25). Hence,
= (0.17, 0.5, 0.33, 0.25) ∧ (0.17, 0.5, 0.33, 0.25) ∧ (0.17, 0.5, 0.33, 0.25) ∧ (0.17, 0.5, 0.33, 0.25)
= (0.07, 0.75, 0.18, 0.07) ∧ (0.17, 0.5, 0.33, 0.25) ∧ (0.17, 0.5, 0.33, 0.25)
= (0.03, 0.88, 0.1, 0.02) ∧ (0.17, 0.5, 0.33, 0.25) = (0.01, 0.94, 0.05, 0.01)
Chapter 5
In subjective logic, regardless of the order how we calculate opinions, we get the same
result as follows.
= ϕ((0.01, 0.94, 0.05, 0.01) ∧ (0, 0, 1, 0.5)) = ϕ(0, 01, 0.94, 0.05, 0.01) = (0, 01, 0.94, 0.05, 0.01)
We believe, above aspect makes subjective logic more solid and sound logic formal-
ism under uncertainty. Especially, compared to fuzzy Belnap logic, the operational order
does not affect on the final result. This is an important aspect, because in fuzzy-Belnap
logic, once we reach at the contradictory point, there is no easy way to escape from the
state unless we use the meet operator ⊗ along the partial order ≤k . Namely, Belnap
149
Contents
logic is non-associative. Therefore, in fuzzy Belnap logic, the sequence of information ar-
rival is important, however, so is not in subjective logic because subjective logic supports
In above setting, given two rules that are considered with the same amount of significance,
we attach more strong belief to f1 . Therefore, the expected result is that the two persons
maybe the same one but not quite certainly. Applying the same inference mechanism
Equation 5.1 for L-fuzzy logics and (2) for subjective logic, the inference results are as
follows. (note that, above setting is not applicable to the case of discrete bilattice species
= [U ⊔ (0.6, 0.3, 0.1) • (0.5, 0.1, 0.4)] ⊕ ¬[U ⊔ (0.3, 0.4, 0.3) • (0.5, 0.1, 0.4)]
150
Contents
= (0.4, 0.07, 0.53) ⊕ ¬(0.24, 0.09, 0.67) = (0.4, 0.07, 0.53) ⊕ (0.09, 0.24, 0.67) = (0.37, 0.21, 0.42)
As shown in Table 5.5, choosing one of t-norm and t-conorm pair and applying Equation 5.8
in Definition 36, we get following inference results derived on L2 , and its interpretations
min / max
clL2 (ϕ)(equal(a, b))
di
= [U ∨ (f1 ∧ r2 )] ⊕ ¬[U ∨ (f2 ∧ r1 )] = [U ∨ (0.6, 0.3) ∧ (0.5, 0.1)] ⊕ ¬[U ∨ (0.3, 0.4) ∧ (0.5, 0.1)]
= [(0, 0)∨(0.5, 0.3)]⊕¬[(0, 0)∨(0.3, 0.4)] = (max(0, 0.5), min(0, 0.3))⊕¬(max(0, 0.3), min(0, 0.4))
= (0.5, 0) ⊕ ¬(0.3, 0) = (0.5, 0) ⊕ (0, 0.3) = (max(0.5, 0), max(0, 0.3)) = (0.5, 0.3)
min / max 2
0.5 0.5·0.3
clIF∗ (L) (clL2 (ϕ)(equal(a, b))) = clIF∗ (L) ((0.5, 0.3)) = ( 0.5+0.3 , 0.5+0.3 ) = (0.31, 0.19)
di
min / max
clIG∗ (L) (clL2 (ϕ)(equal(a, b))) = clIG∗ (L) ((0.5, 0.3)) = (0.5 − 0.3 0.3
2 , 2 ) = (0.35, 0.15)
di
Chapter 5
prod/sum
clL2 (ϕ)(equal(a, b))
di
= [U ∨ (f1 ∧ r2 )] ⊕ ¬[U ∨ (f2 ∧ r1 )] = [U ∨ (0.6, 0.3) ∧ (0.5, 0.1)] ⊕ ¬[U ∨ (0.3, 0.4) ∧ (0.5, 0.1)]
= [U ∨ (0.6 · 0.5, 0.3 + 0.1 − 0.3 · 0.1)] ⊕ ¬[U ∨ (0.3 · 0.5, 0.4 + 0.1 − 0.4 · 0.1)]
= (0.3, 0) ⊕ ¬(0.15, 0) = (0.3, 0) ⊕ (0, 0.15) = (0.3 + 0 − 0.3 · 0, 0 + 0.15 − 0 · 0.15) = (0.3, 0.15)
prod/sum 2
0.3 0.3·0.15
clIF∗ (L) (clL2 (ϕ)(equal(a, b))) = clIF∗ (L) ((0.3, 0.15)) = ( 0.3+0.15 , 0.3+0.15 ) = (0.2, 0.1)
di
prod/sum
clIG∗ (L) (clL2 (ϕ)(equal(a, b))) = clIG∗ (L) ((0.3, 0.15)) = (0.3 − 0.15 0.15
2 , 2 ) = (0.225, 0.075)
di
Luk
clL 2 (ϕ)(equal(a, b))
di
= [U ∨ (f1 ∧ r2 )] ⊕ ¬[U ∨ (f2 ∧ r1 )] = [U ∨ (0.6, 0.3) ∧ (0.5, 0.1)] ⊕ ¬[U ∨ (0.3, 0.4) ∧ (0.5, 0.1)]
= [U ∨ (max(0, 0.6 + 0.5 − 1), min(0.3 + 0.1, 1)] ⊕ ¬[U ∨ (max(0, 0.3 + 0.5 − 1), min(0.4 + 0.1, 1)]
= [(0, 0) ∨ (0.1, 0.4)] ⊕ ¬[(0, 0) ∨ (0, 0.5)] = (min(0 + 0.1, 1), max(0, 0 + 0.4 − 1))
151
Contents
drastic
clL2 (ϕ)(equal(a, b))
di
= [U ∨ (f1 ∧ r2 )] ⊕ ¬[U ∨ (f2 ∧ r1 )] = [U ∨ (0.6, 0.3) ∧ (0.5, 0.1)] ⊕ ¬[U ∨ (0.3, 0.4) ∧ (0.5, 0.1)]
= [U ∨ (TD (0.6, 0.5), SD (0.3, 0.1)] ⊕ ¬[U ∨ (TD (0.3, 0.5), SD (0.4, 0.1)]
= [(0, 0) ∨ (0, 1)] ⊕ ¬[(0, 0) ∨ (0, 1)] = (SD (0, 0), TD (0, 1)) ⊕ ¬(SD (0, 0), TD (0, 1))
= (0, 0) ⊕ ¬(0, 0) = (0, 0) ⊕ (0, 0) = (SD (0, 0), SD (0, 0)) = (0, 0)
N ilpotent
clL2 (ϕ)(equal(a, b))
di
= [U ∨ (f1 ∧ r2 )] ⊕ ¬[U ∨ (f2 ∧ r1 )] = [U ∨ (0.6, 0.3) ∧ (0.5, 0.1)] ⊕ ¬[U ∨ (0.3, 0.4) ∧ (0.5, 0.1)]
= [U ∨ (TnM (0.6, 0.5), SnM (0.3, 0.1)] ⊕ ¬[U ∨ (TnM (0.3, 0.5), SnM (0.4, 0.1)]
= [(0, 0)∨(0.5, 0.3)]⊕¬[(0, 0)∨(0, 0.4)] = (SnM (0, 0.5), TnM (0, 0.3))⊕¬(SnM (0, 0), TnM (0, 0.4))
= (0.5, 0) ⊕ ¬(0, 0) = (0.5, 0) ⊕ (0, 0) = (SnM (0.5, 0), SnM (0, 0)) = (0.5, 0)
N ilpotent 2
0.5 0.5·0
clIF∗ (L) (clL2 (ϕ)(equal(a, b))) = clIF∗ (L) ((0.5, 0)) = ( 0.5+0 , 0.5+0 ) = (0.5, 0)
di
N ilpotent
clIG∗ (L) (clL2 (ϕ)(equal(a, b))) = clIG∗ (L) ((0.5, 0)) = (0.5 − 02 , 20 ) = (0.5, 0)
di
Hamacher
clL2 (ϕ)(equal(a, b))
di
= [U ∨ (f1 ∧ r2 )] ⊕ ¬[U ∨ (f2 ∧ r1 )] = [U ∨ (0.6, 0.3) ∧ (0.5, 0.1)] ⊕ ¬[U ∨ (0.3, 0.4) ∧ (0.5, 0.1)]
= [U ∨ (TH0 (0.6, 0.5), SH2 (0.3, 0.1)] ⊕ ¬[U ∨ (TH0 (0.3, 0.5), SH2 (0.4, 0.1)]
= [(0, 0) ∨ (0.375, 0.39)] ⊕ ¬[(0, 0) ∨ (0.23, 0.48)] = (SH2 (0, 0.375), TH0 (0, 0.39))
⊕ ¬(SH2 (0, 0.23), TH0 (0, 0.48)) = (0.375, 0) ⊕ ¬(0.23, 0) = (0.375, 0) ⊕ (0, 0.23)
Figure 5.12 shows above results in opinion space, I ∗ (L) and L2 . Unlike the case
of using values lying on the boundary of the spaces such as T, F, DT1 , DF1 , when internal
values in spaces are used, the reasoning results are quite dependent on the choice of t-norms
152
Contents
Figure 5.12: Reasoning Results of Modified Example 3 in Opinion Space and L2 & I ∗ (L).
Chapter 5
and t-conorms. However, what pair of t-norms and t-conorms to use is not easy to answer.
This is a problem common to all fuzzy set based applications. Typically, connectives are
categorized by the properties they satisfy. Lukasewicz connectives are in some sense the
most interesting one because they satisfy the most properties of binary connectives, but
it does not mean that they are best suited for each application. This aspect is some-
times also attacked by statisticians who prefer Bayesian theory. However, Product / Sum
connectives are interesting in Bayesian sense, because Product t-norm and Sum t-conorm
resemble probabilistic conjunction and disjunction. For instance, following Equation 5.8,
fuzzy Belnap connectives on L2 that are compositionally defined upon Product t-norm and
153
Contents
As mentioned throughout this chapter, subjective logic has solid mathematical basis in
Bayesian perspective on dealing binary (crisp) event (see preliminaries and definitions oper-
ators). Therefore, it is worth to compare above Product / Sum connectives with the ones in
subjective logic in the definition level. For example, given two opinions wx = (bx , dx , ix , ax )
dx∧sl y = dx + dy − dx dy
(1−ay )bx iy +(1−ax )ix by
ix∧sl y = ix iy + 1−ax ay
ax∧sl y = ax ay
and the disjunction of the two opinions are defined as follows (see Definition 32):
bx∨sl y = bx + by − bx by
ax (1−ay )dx iy +(1−ax )ay ix dy
dx∨sl y = dx dy + ax +ay −ax ay
ay dx iy +ax ix dy
ix∨sl y = ix iy + ax +ay −ax ay
ax∨sl y = ax + ay − ax ay
the consensus ⊗sl of the two opinions are defined as follows (see Definition 30):
bA,B
x = (bA B B A
x ix + bx ix )/k
dA,B
x = (dA B B A
x ix + dx ix )/k
iA,B
x = (iA B
x ix )/k
x ix +ax ix −(ax +ax )ix ix
aA A B A A B A B
aA,B
x = iA +i B −2iA iB
x x x x
x + ix − ix ix .
where, k = iA B A B
logic look similar, they are not exactly the same. The reason is because the definition in
subjective logic is defined so that it can model a beta distribution that approximates the
tions of the given two opinions wx and wy [97] (note that, the result of multiplication and
comultiplication of two beta functions are not always beta function, [97]).
Similarly, while the join operator ⊕bl on Product / Sum just sum both the belief and
the disbelief, subjective logic calculation is designed so that it can model the beta distribu-
154
Contents
tion derived by merging each pair of parameters of the beta distributions correspond to the
given two opinions wx and wy [94]. Indeed, through (5.2), (5.3) and (5.4) in Section 5.6.3,
we have shown that the ‘consensus’ operator itself is the calculation of ‘Bayes Theorem’
itself. Due to this aspect, even compared with Product / Sum fuzzy Belnap connectives,
When it comes to visual surveillance, number of vision analytics are based on the
pattern recognition and machine learning techniques that are also (in many cases) based
on Bayesian statistics rather than fuzzy theory. Noting this aspect, we advocate subjective
logic could be better suited for visual surveillance applications especially when we want to
stay closer to the way that usual vision analytics generate uncertain symbolic facts.
In summary of lessons learned in this chapter, the attempt to modeling default reasoning
Chapter 5
using subjective logic has given several insights. 1) It shows the feasibility of proposed
subjective logic have commonality in representing epistemic status, subjective logic is more
close to Bayesian on operation of such status. 3) The logical soundness of the proposed
approach makes it attractive for visual surveillance scenarios compared with L-fuzzy logic
In this chapter, we proposed subjective logic based inference framework for default
reasoning, and demonstrated its use for high level semantic analysis of visual surveillance
under incomplete and imprecise knowledge, that can play an important role for deriving
plausible conclusions for many applications. Especially, in the forensic sense of visual
ter an incident or a report, it is natural to examine all positive and negative contextual
155
Contents
evidences that are related to the given hypothesis and fuse them to derive plausible conclu-
sion based on default reasoning. The keys to enable default reasoning are 1) representing
due to its property of representing belief with ignorance and its rich set of operators for
handling uncertain beliefs. To contrast the properties and advantage of the proposed ap-
proach, we also applied the inference scheme on L-fuzzy set based logics. The case study
results show that the proposed approach and L-fuzzy set based approaches can be an al-
ternative tool to model default reasoning. Among the L-fuzzy logics, intuitionistic fuzzy
logic is very similar to the uncertainty representation scheme of subjective logic. While
the generalized intuitionistic fuzzy logic, that is fuzzy Belnap logic, can be defined on a
bilattice structure with operators regarding degree of information, intuitionistic fuzzy logic
could not be fully defined on a bilattice structure because the join operator along the axis
of degree of information can not be defined. Contrary to intuitionistic fuzzy logic, even
though it also has triangle structure, subjective logic provides an operator called consensus
that has very similar behaviour as the join operator on degree of information in bilattice.
This is because when two opinions are fused by the consensus operator, it always decreases
ignorance in the derived opinion except in the case of fusing definite true (full belief) and
definite false (full disbelief). Due to this aspect, the comparison of subjective logic based
default reasoning with intuitionistic fuzzy logic was done via a mapping between the fuzzy
Belnap logic and the intuitionistic fuzzy logic. The reasoning result of both the subjective
logic and the fuzzy Belnap logic seem reasonable. However, as noted in the discussion
section, fuzzy Belnap logic has some problems. 1) the truth table has some problematic
aspects, thereby logically not sound. 2) due to 1) the sequence of getting information is
critical. 3) due to 1) once the epistemic state is reached to the contradictory state, it is
not easy to escape that state. 4) the basic four logical operators in L-fuzzy logics can be
determined in many ways, therefore, the semantics of the operators are not sound and clear
in Bayesian sense. Due to these aspects, we advocate subjective logic has advantages as a
156
Contents
tool for artificial reasoning in visual surveillance. Because, in visual surveillance, due to the
flexibility, and instability of the vision analytics, we can not guarantee the sequence of get-
ting information, therefore, the reasoning system should be robust against the information
acquisition sequence. Indeed, most of the vision analytics are based on probabilistic theory,
therefore, the values from those analytic modules could be well interpreted in subjective
logic. Beside these aspects, there is yet another advantage of the proposed approach, that
is the ability of default reasoning can be fulfilled within a single subjective logic based
reasoning framework that can also offer additional potential usage such as bidirectional
conditional modeling [81], reputation based belief decaying, etc. [80]. Therefore, enabling
default reasoning to subjective logic could offer better expressive power for modeling and
There are, however, still open issues such as comparing the introduced inference
scheme to more complicated situational reasoning. Therefore, our future research will
cover such comparisons and applying the shown approach to more complicated scenarios
Chapter 5
using automatically generated large scale data.
157
158
6 Vague Proposition Modeling and Handling
In the previous chapter, we have explained that subjective logic can be a good means
to model ‘default reasoning’. The main idea was to label truth of facts and rules with
subjective opinion (see Definition 33) and to fuse both positive and negative sources of
Section 4.3.2, the former corresponds to 1 - 2 of Definition 16 and the latter corresponds
logic calculation scheme written in the rule body. In this chapter, we will further explore
Chapter 6
6.1 Introduction
This chapter presents an approach to modeling vague contextual rules using subjective logic
for forensic visual surveillance. Unlike traditional real-time visual surveillance, forensic
analysis of visual surveillance data requires mating of high level contextual cues with
observed evidential metadata where both the specification of the context and the metadata
suffer from uncertainties. To address this aspect, in Chapter 4, we proposed the use of
declarative logic programming to represent and reason about contextual knowledge, and
the use of subjective logic for uncertainty handling. Upon this approach, in the previous
Chapter 5, we have demonstrated that subjective logic can be a good means to model
159
Contents
‘default reasoning’. The main idea was to label generalized truth or priority of facts and
rules with subjective opinion and to fuse both positive and negative sources of information
together.
However, there are often cases that the truth value of rule itself is also uncertain
thereby, uncertainty assignment of the rule itself should be rather functional. ‘the more X
then the more Y’ type of knowledge is one of the examples. To enable such type of rule
modeling, in this chapter, we propose a reputational subjective opinion function upon logic
programming, which is similar to fuzzy membership function but can also take uncertainty
of membership value itself into account. Then we further adopt subjective logic’s fusion
operator to accumulate the acquired opinions over time. To verify our approach, we present
that uses metadata extracted by a person tracker and evaluates the relationship between
the tracked persons. The case study is further extended to demonstrate more complex
In the pipeline of our proposed approach (see Chapter 4), intermediate metadata comes
from vision analytics and additional visual or non visual contextual cues are encoded as
either symbolized facts or rules. Then uncertainty comes with vision analytics are repre-
sented as subjective opinions and attached to their symbolized facts. Similarly, uncertainty
as general trustworthiness or priority among rules is also represented and attached to given
contextual rules. Once such uncertainty attachment is done, principled inference, which
is often nonmonotonic, is conducted. The examples of such principled inferences are de-
fault reasoning [143] to handle inconsistent information (see Chapter 5), abduction [50] to
find most probable hypothesis of given observation and belief revision over time upon the
change of observation (see Chapter 8), etc. Therefore, appropriate uncertainty assignment
160
Contents
Unlike above mentioned opinion assignment to logical rules, however in reality, there
are linguistically vague logical rules that hinders assigning a single opinion to a rule does
not make sense but opinion assignment itself should be done rather functionally. In this
chapter we will refer to this type of rule as ‘Vague Rule’. An example rule in this type
is ‘the more X then the more Y’. Humans are very skillful in dealing such type of rules.
To enable such type of rule handling in our framework, we first examine a subjective logic
function that is similar to fuzzy membership function but can also take uncertainty of
membership value itself into consideration. To demonstrate reasoning under such vague
type of available metadata to the results from human detection and tracking algorithms.
Automatic human detection and tracking is one of the common analytics and becoming
more widely employed in automated visual surveillance systems. The typical types of meta-
information that most human detection analytic modules generate comprise, for instance,
localization information such as coordinate, width, height, time and (optionally) additional
low-level visual feature vectors. We intend to use further such information for evaluating
the relationship between two persons and, more specifically, for estimating whether one
Chapter 6
person could serve as a witness of another person in a public area scene. Examples for
(linguistic) domain knowledge applicable to this scenario include: 1) (At least) two distinct
people are required for building a relationship. 2) The closer the distance between two
people is, the higher is the chance that they may identify each other. 3) If two persons
approach each other directly (face-to-face) then there is a higher chance that they can
identify each other. Such linguistic knowledge can be modeled and encoded as rules by
the proposed approach. The case study is further extended to demonstrate more complex
forensic reasoning by considering additional contextual rules together with the shown vague
rules.
161
Contents
The proposed vague rule modeling approach mainly relies on the logic programming ex-
tended with subjective logic (see Section 4.3.2). Firstly, for a given propositional know-
ledge, we assume a fuzzy-like membership function that grades degree of truth. Then we
focus on that the interpretation of such membership function can be dogmatic, thereby,
when the function is projected on the opinion space, it only lays on the bottom line of the
opinion space. Indeed, in many cases, the exact shape of the function is hard to deter-
mine. To address this aspect, we introduce a reputational function that evaluate the trust
However, there are often cases that we may want to interpret the meaning as ‘the more x
then the more y’ or ‘the more x then the less y’, etc. In this case, the opinion attached to
the consequence of the rule should be rather functional in terms of the elements within the
rule body. Therefore, the opinion assignment suit to this interpretation is Definition 16.
In the sense of intrinsic linguistic uncertainty of the rule, it resembles fuzzy rules shown by
Anderson et al. [17, 177]. In the work, quantitative low level features of human detection
results such as ‘centroid’, ‘eigen-based height’ and ‘ground plane normal similarity’ are lin-
guistically mapped into non-crisp premises (i.e. fuzzy sets) as ‘(H)igh’, ‘(M)edium’, ‘(L)ow’
and ’(V)ery Low’. Then fuzzy rules defines the conjunctive combination of those linguistic
symbols to draw higher semantics such as ‘Upright’, ‘In Between’ and ‘On the ground’
introducing appropriate fuzzy membership functions for each linguistic terms and proper
handling of the membership functions is of important issue. In this view, Mizumoto et al.
162
Contents
such as ‘more or less’ having various shapes. One another thing worth to note concern-
ing Fuzzy Logic is that, even if there are Zadeh’s original logical operators, there are yet
another ways of defining logical operators as well. For example, for given two quantita-
tive variables x and y come with corresponding membership functions µa and µb , Zadeh’s
AND operator is defined as ‘x AND y = min(µa (x), µa (y))’. In socalled ‘t-norm fuzzy
logic’, any form of t-norms can be considered as AND operators (see Table 5.5). (note
that, t-norms also played an important role to define lattice operators of L-fuzzy sets as
shown in Definition 5.8 of Chapter 5). For example, in the case of using product t-norm,
the AND operator can be defined as ‘x AND y = µa (x) · µb (x)’ [73]. This aspect still
remains controversial among most statisticians, who prefer Bayesian logic [179]. Contrary,
as explained in the section Section 3.1, subjective logic can be interpreted in the sense of
bayesian and also the final quantitative opinion space can also be interpreted in the sense
of fuzziness (i.e. ‘very certainly true’, ‘less certainly true’, etc). This way, we believe that
subjective logic can better bridges the interpretation of fuzzy intuitive concepts in bayesian
• 1. For a given propositional rule ‘the less (more) y ← the more x’ we could introduce a
membership-like function µi : x → y.
Chapter 6
• 2. It is clear that the function µi should be monotonically decreasing (increasing) but
values of µi (x) at the two extreme point of (minx ≤ x ≤ maxx ) tend to converge but
the values in between are diverge therefore, the values of later cases are more uncertain.
the function µi and combine it with raw opinion obtained from µi using subjective logic’s
reputation operator.
163
Contents
Figure 6.1: Vague Rule Modeling using Subjective Logic’s Reputation Operator.
This idea is depicted in Figure 6.1, where the actual reputational operation is defined as
follows :
Definition 41. (Reputation) [99] . Let A and B be two agents where A’s opinion about
A = {bA , dA , uA , aA }, and let x be a proposition
B’s recommendations is expressed as wB B B B B
where B’s opinion about x is recommended to A with the opinion wxB = {bB
x , dx , ux , ax }.
B B B
then wxA:B is called the reputation opinion of A. By using the symbole ⊗ to designate this
A ⊗ wB .
operation, we get wxA:B = wB x
given µi need to be defined. Although there could be also another ways of such function,
164
Contents
bx = k + 4(1 − k)(µi (x) − 12 )2
µi (x)
wµreput = 1−bx
dx = Dratio (6.1)
i (x)
ux = 1 − bx − dx
where k, represents the minimum boundary of belief about the value from µi (x), and
the Dratio indicates the ratio for assigning the residue of the value µi to disbelief and
At this stage we focused on evaluating the modeling approach itself rather than the reli-
ability of the person detection algorithm. Therefore, we manually annotated a test video
from one of i-LIDS [3] data sample with ground truth metadata for human detection com-
prising bounding boxes and timing information (shown in Figure 6.2). In total, 1 minute
of test video was annotated in which there are 6 people. For our purposes, we intentionally
marked one person as suspect. Then we encoded following linguistic contextual knowledge
according to the proposed approach as explained in Section 4.3.2. 1) (At least) two dis-
Chapter 6
tinct people are required for building a relationship 2) The closer the distance between
two people is, the higher is the chance that they can identify each other. 3) If two persons
approach each other directly (face-to-face) then there is a higher chance that they can
identify each other. Then we calculate subjective opinions between the person marked as
6.4.2.1 Distance
The distance between a pair of people would be one of the typical pieces of clue for rea-
soning whether one person could serve as a witness of another person. This relates to the
165
Contents
general human knowledge that ‘The closer two people are in distance, the more chances of
perceiving the other are’. Humans are very adapted to operating upon such type of uncer-
tain and ambiguous knowledge. Exactly modeling such a relation is not trivial, but we can
each other. This aspect is depicted as three possible curves in the middle of Figure 6.3
- A.), where x represents the distance between the persons as calculated from the person
detection metadata and µi represents the likelihood that two persons at this distance would
perceive each other, maxdist is the maximum possible (i.e diagonal) distance in a frame
and ai is the estimated probability that two humans could’ve recognized each other at the
maxdist distance. However, the value derived from such function is not fully reliable due
to the variety of real world and uncertainty in the correctness of the function and uncer-
tainty in the distance value itself. Considering the aspect of distance, it is clear that both
166
Contents
extreme cases i.e. very close or very far are much more certain than in the middle of the
range. Thus, to better model the real world situation, the reputational opinion function
need to be applied to any chosen function µi . This is modeled as opinion on the reliability
of µi (x) by applying Equation 6.1. In order to evaluate the impact of choosing different
functions in Figure 6.3 - A.), three different types of µi functions (a concave, convex and
linear) have been applied. The derived reputational opinions showed similar aspects having
6.4.2.2 Direction
Similarly, we also used direction information between two persons. The linguistic know-
ledge to be modeled is ‘if two persons approach each other directly (face-to-face) then there
function is shown in Figure 6.3 - B.), where Θ represents the angle between the persons
heading directions as calculated from the person detection metadata and µi represents the
likelihood that two persons at the angle would perceive each other and ai is the expected
minimum probability that two humans could’ve recognized each other at any angle. How-
ever, again the trustworthiness of the values from such functions µi is uncertain, especially
Chapter 6
in the middle range of the Θ. To roughly model such aspect, for a chosen function µi (Θ),
the same reputational function from Equation 6.1 was used again. The impact of choosing
different µi showed similar behavior as of direction based opinions as shown in Figure 6.4.
167
Contents
Figure 6.4: Samples of Reputational Opinion according to Distance and Equation 6.1.
168
Contents
In addition to the uncertainty modeling, logic programming is used to represent the given
contextual rules as explained in Section 4.3.2. Encoded rules in form of Equation 4.2 are
as follows :
Rule1
wwitness(H 1 ,H2 ,T1 )
←
µ( d) (6.2)
HumanD etector HumanD etector µ (d)
(whuman(H1 ,T1 )
∧ whuman(H 2 ,T1 )
) ⊗ (wwitness(H
dist
1 ,H2 ,T1 )
⊗ wµreput
dist (d)
)
Rule2
wwitness(H1 ,H2 ,T1 )
←
µ(d)
(6.3)
HumanD etector HumanD etector µ (d)
(whuman(H1 ,T1 )
∧ whuman(H 2 ,T1 )
) ⊗ (wwitness(H
dir
1 ,H2 ,T1 )
⊗ wµreput
dir (d)
)
Rule3
wwitness(H1 ,H2 ,T1 )
← (wwitness(H
Rule1
1 ,H2 ,T1 )
∧ wwitness(H
Rule2
1 ,H2 ,T1 )
) (6.4)
Rule4
wwitness(H 1 ,H2 ,Tn )
← ⊕n Rule3
i=1 wwitness(H1 ,H2 ,Ti ) (6.5)
The first Rule 6.2 starts considering the necessary condition, meaning that there should
be a distinct pair of two people. Therefore the conjunction operation ∧ (see Definition 31)
on two opinions [97] is used that is very similar to the operation P (A) · P (B) except
that in subjective logic the opinion can additionally represent ignorance. Then, for the
resulting set of frames the reputational opinion about the distance opinions is calculated
as described in Section 6.4.2. Each result is assigned to a new opinion with the predicate
Chapter 6
of the appropriate arity and is assigned the name of agent with the final belief values. In
this case, the final opinion value represents that there is an opinion about two persons
being potential witnesses of each other from an agent named Rule1. The second Rule 6.3
is almost same as Rule 6.2. The only different part of this rule is that the reputational
opinion is about direction. The third Rule 6.4 combines the evidences coming from Rule 6.2
and (6.3). The conjunction operator ∧ is used to reflect that for reliable positive resulting
opinions both evidences should have appeared with a certain amount of belief. The last
Rule 6.5 is about accumulating the belief over time using the consensus operator ⊕ [94]
(see Definition 30). Figure 6.5 shows a graphical representation of the rules in a tree form.
169
Contents
Using the rules described in Section 6.4.3, we calculated subjective opinions between a per-
son marked as suspect and other human instances over time. Figure 6.6 shows a snapshot
of the visualization in the prototype comprising a video player and an opinion visualizer.
While the video is being played the corresponding metadata is transformed into the cor-
responding opinion representation. The translated opinions are fed into the rule-engine
170
Contents
which automatically evaluates the rules. The right part of Figure 6.6 shows the opinion
about the proposition ‘human 5 is a witness for the suspects’ marked red and its corre-
was prepared to collect scores about the witnessing chances for each of the ‘pairs’ in the
scene (e.g. human1 and suspect, human2 and suspect , etc). 7 people from our lab took
part in the questionnaire. Then changing the uncertainty functions on vague rules, we
tested the behavior of the proposed approach to check whether it well models human in-
distance functions and 3 direction functions), to better contrast the impact of changing
such uncertainty functions, we have fixed the direction function to the type of µ3 defined
Chapter 6
in Figure 6.3 - B.) and tested with 3 different direction functions shown in Figure 6.3 -
A.). Then the mean and standard deviation, min and max of the ‘human opinions’ were
calculated and compared to the computed results. According to [93], the following criteria
1) The opinion with the greatest probability expectation is the greatest opinion.
3) The opinion with the least relative atomicity is the greatest opinion.
In the described experiment, due to the small size of possible pairs, only the first cri-
terion was applied and the final expectation values of each opinion for candidate pairs
were plotted jointly with the questionnaire based result as shown in Figure 6.7. The final
result turns out to be following the tendency of questionnaire based human ‘opinions’. The
171
Contents
change of uncertainty function seems not introducing that critical differences. However,
there were more differences between the expected values, when the final expectation values
were low, for instance, though it was a slight differences, µ3 tend to yield larger expecta-
tion value then µ2 and µ1 . The differences ware smaller when the final expectation values
were getting higher. However, in any cases, the order on the ranking of witnesses show the
same results. Therefore, in the sense of human like reasoning, it seems that the proposed
In this section, we further explorer the proposed case study scenario for more complex
contextual forensic reasoning. Especially, we will consider the situation that is needed to
Let us consider a conceptual scenario that a security personnel wants to get sugges-
172
Contents
that automatic vision analytics are running and extracting basic semantics, we will assume
two virtual situations as shown in Figure 6.8, where, witnesses are reasoned according to
will assume that ‘witness2’ has higher opinion then ‘witness1’. In addition to this, we will
assume optional cases that additional evidential cues are detected. In Figure 6.8 - A.),
‘witness2’ is talking on the phone. In Figure 6.8 - B.), the optional case is the detection of
a license plate of the car seems to belong to the ‘witness1’ and ‘witness2’ comes with face
detection.
Given the scenario with optional cases, we will also assume that 1). people usually don’t
recognize well when they are talking on the phone, 2). identifiable witness is a good wit-
ness. 3) License plate is better identifiable source than face detection because we can even
fetch personal information of the owner easily. Therefore, under optional assumption, for
example, in Figure 6.8 - A.), ‘witness1’ should be better witness, and in Figure 6.8 - B.),
‘witness1’ should be suggested as a better witness. This kind of non monotonic reasoning
under inconsistent information can also be regarded as ‘default reasoning’ [143]. In Chap-
Chapter 6
ter 5, we showed that this aspect can be modeled using subjective logic as well under the
opinion assignment (see Definition 33) and inference mechanism (see Definition 35) shown
in Chapter 5. Here, it is important to note that, unlike the case of vague rule modeling,
the type of opinion assignment to prioritize belong to Definition 16 - 2. and the default
(full truth), DT1 ≃ (0.5, 0, 0.5) (weak default true ), DT2 ≃ (0.8, 0, 0.2) (strong default
true), F ≃ (0, 1, 0) (full false), DF1 ≃ (0, 0.5, 0.5) (weak default false), DF2 ≃ (0, 0.8.0.2)
(strong default false), ∗ ≃ (0.33, 0.33, 0, 34) (contradiction), U ≃ (0, 0, 1) (full uncertainty)
and ⊥ ≃ (0.5, 0.5, 0) (full contradiction) [82, 83]. For the rest of truth values we will use
173
Contents
Example 6. (Witness talking on the phone) . Assume the following set of rules
about determining good witness including the uncertain spatio-temporal relation based
witness reasoning rule described in Section 6.4.3. Then also assume the following opinion
assignment that witness2 (denoted as wit_2) has higher opinion being the witness than
the inference for reasoning better witness using default logic with subjective logic is as
follows.
clsldi (ϕ)(wwitness(wit_1) ) = [U ⊔ ((0.6, 0.15, 0.25) · DT1 )]
= [U ⊔ (0.44, 0.15, 0.41)] = (0.44, 0.15, 0.41) ∼ (Expectation = 0.54)
clsldi (ϕ)(wwitness(wit_2) ) = [U ⊔ ((0.7, 0.10, 0.20) · DT1 )]
= [U ⊔ (0.50, 0.10, 0.40)] = (0.50, 0.10, 0.40) ∼ (Expectation = 0.60)
Above result shows that given the weak rules, ‘witness2’ is more probable witness candidate
than ‘witness1’. Then, let us consider the weak opinion assignment to the additional
contextual cue that witness2 is using the phone. This semantics can be interpreted as ‘the
Given the additional information, the inference on witness2 is being witness is as follows.
clsldi (ϕ)(wwitness(wit_2) )
= [U ⊔ ((0.7, 0.10, 0.20) · DT1 )] ⊕ ¬[U ⊔ ((0.6, 0.15, 0.25) · DT2 )]
= [U ⊔ (0.50, 0.10, 0.40)] ⊕ ¬[U ⊔ (0.59, 0.15, 0.26)]
= (0.50, 0.10, 0.40) ⊕ ¬(0.59, 0.15, 0.26)
= (0.50, 0.10, 0.40) ⊕ (0.15, 0.59, 0.26)
= (0.34, 0.47, 0.19) ∼ (Expectation = 0.39)
The resulting opinion (0.34,0.47,0.19) on witness2’s being a good witness now weaker than
(0.44,0.15,0.41) which is for the case of witness1’s being a good witness. The expectation
174
Contents
values also captures this aspect. Thus, this result shows that the inference scheme well
Example 7. (Witness with Face Detection vs. License Plate Detection) . Con-
sider the following set of rules about determining good witness and the following opinion
assignment to capture the scenario described in Section 6.5.1 as depicted in Figure 6.8.
Rule4
ϕRule [wwitness(H1)
← ⊕ni=1 wwitness(H
Rule3
1 ,H2 ,Ti )
] = DT1
ϕRule [wwitness(H1 ) ← wwitness(H
Rule4
1)
· whasF aceDetectInf o(H1 ) ] = DT1
ϕRule [wwitness(H1 ) ← wwitness(H
Rule4
1)
· whasLicenseDetectInf o(H1 ) ] = DT2
Rule4
ϕRuleEval [wwitness(wit_1) ] = (0.6, 0.15, 0.25)
Rule4
ϕRuleEval [wwitness(wit_2) ] = (0.7, 0.10, 0.20)
ϕf act [whasLicenseDetectInf o(wit_1) ] = (0.6, 0.15, 0.25)
ϕf act [whasF aceDetectInf o(wit_2) ] = (0.6, 0.15, 0.25)
Given two default true and default false rules and facts that can be seen as definite true,
the inference for reasoning better witness using default logic with subjective logic is as
follows.
clsldi (ϕ)(wwitness(wit_1) ) = [U ⊔ ((0.6, 0.15, 0.25) · DT1 · (0.6, 0.15, 0.25) · DT2 )]
Chapter 6
clsldi (ϕ)(wwitness(wit_2) ) = [U ⊔ ((0.7, 0.10, 0.20) · DT1 · (0.6, 0.15, 0.25) · DT1 )]
Above result shows that given the evidences, ‘witness2’ is slightly more probable witness
candidate than ‘witness1’ because license plate info is more informative thereby strongly
considered than face related information by the opinion assignment. However, due to the
opinion on the fact level is not certain, the values were not strongly forced the belief but
rather increased the uncertainty in the final opinion. The expectation values also captures
this aspect. Thus, this result show that the inference scheme well models human intuition.
175
Contents
assessment of a vague proposition offers more choices to model complex contextual human
knowledge by enriching the expressive power of the framework. The proposed approach
can be used with another principled reasoning scheme such as default reasoning. There
are, however, still open issues on automatic assignment of proper priors and proper
Intelligent forensic reasoning upon metadata acquired from automated vision ana-
lytic modules is an important aspect of surveillance systems with high usage potential.
The knowledge expressive power of the reasoning framework and the ability of uncertainty
handling are critical issues in such systems. In this chapter, based on our previous work
on the use of logic programming with subjective logic, we extended the framework so that
it can also handle vague propositional rules. The approach is mainly based on the fuzzy-
like membership function and the reputational operation on it. The main advantage of
the proposed approach is that it offers more choices to model complex contextual human
knowledge by enriching the expressive power of the framework. The other advantage of the
proposed approach is that the modeled vague rules can be used with another principled
reasoning scheme. In this chapter, especially, we have demonstrated how the reasoning re-
sults from uncertain spatio-temporal rules could be used with default reasoning. Another
interesting properties of the system is that, unlike traditional probability based conditional
reasoning, this approach allows for representing lack of information about a proposition.
We could also roughly assign our subjective priors with lack of information, and obser-
vations can also be represented with any degree of ignorance, therefore we believe this
better reflects human intuition and real world situations. Another beneficial property is
the flexibility of assigning opinions to formulae. Especially, rule can embed its own opinion
calculation scheme thereby, allows for sophisticated propagation of opinions through the
inference pipeline. There are, however, still several open issues such as how to better model
176
Contents
the reputational function, how to automatically assign proper prior opinions to rules, etc.
Although we still need to extend this concept to large scale data. We advocate that this
Chapter 6
177
178
7 Hybrid Knowledge Modeling and Reasoning
Thus far, our discussions have been focused on the ‘extensional’ aspect of knowledge rep-
approach has benefits on ‘flexibility’ and ‘expressive power’ due to the ‘modularity’, it also
has some deficiencies such as ‘improper handling of bidirectional inference’ that can be
benefits of both approaches in a single framework. In this chapter, we focus on the exten-
sion of our proposed reasoning framework to bestow ‘intensional’ characteristics upon our
7.1 Introduction
inference under uncertainty play an important role for deriving new contextual cues by
fusing relevant evidential patterns. To address this aspect, both rule-based (aka. exten-
sional) and state based (aka. intensional) approaches have been adopted for situation or
visual event analysis. The former provides flexible expressive power and computational
Chapter 7
efficiency but typically allows only one directional inference. The latter is computationally
consequent of conditionals as mutually relevant states (see Section 2.4 for details). In vi-
sual surveillance, considering the varying semantics and potentially ambiguous causality in
conditionals, it would be useful to combine the expressive power of rule-based system with
179
Contents
that, while relying mainly on a rule-based architecture, also provides an intensional way of
show how conditionals can be assessed via explicit representation of ignorance in subjective
logic. We then describe the proposed hybrid conditional handling framework. Finally, we
present several experimental case studies from a typical public domain visual surveillance
scenes.
In recent years, there has been an increasing research focus on higher level semantic reason-
ing for visual surveillance data by augmenting low level computer vision modules with high
level contextual cues (see Section 1.1 for detailed background on this). Considering the
events, intelligent high level semantic reasoning should provide a means of fusing evidence
The key challenges for such high level reasoning approaches are the choice of an appropriate
As we have reviewed in Section 2.3 and Section 2.4, depending on how such ap-
proaches handle uncertainty, they can be roughly categorized into ‘extensional’ and ‘inten-
sional’ approaches [130]. Extensional approaches also known as rule-based systems treat
uncertainty as a generalized truth value attached to formulas and compute the uncer-
tainty of any formula as a function of the uncertainties of its sub formulas. In intensional
approaches, also known as state based approaches, antecedents and consequents in condi-
tionals are treated as ‘subsets of possible states’ and handle uncertainty taking into account
relevance between the two states. Extensional approaches have advantages in the ‘flexibil-
ity’ and ‘expressive powe’ due to their ability to derive a new proposition based only on
what is currently known (a) regardless of anything else in the knowledge base (‘locality’)
and (b) regardless of how the current knowledge was derived (‘detachment’). ‘locality and
180
Contents
intensional approaches. This deficit of semantic clarity comes from how conditionals are
not possible. Although it is possible to explicitly add reverse conditionals, this would in-
troduce cycles that can adversely impact the reasoning mechanism [130]. Another problem
of conditional formulas. Due to the inference direction and modularity, the uncertainty
propagation is limited to one direction and, thus hinders considering relations between an-
reasoning that focusses on the probabilistic relation between antecedent and consequent.
possible worlds via set theory operations. Such approaches exhibit neither the property of
locality nor detachment. In other words, they are unable to derive new information unless
they have exhaustively accounted for all other information sources in the knowledge base,
that could possibly influence the final proposition to be reasoned about. While this process
provides better semantic clarity, it comes with an exponential computational cost. Another
Since intensional approaches possess complete knowledge of all information sources influ-
encing all variables of interest, it is possible to ask arbitrary queries of the model without
be related but the causality direction is ambiguous due to the complexity and variety of
semantics. Thus, it would be desirable to combine the two approaches to achieve more
context modeling power and computational efficiency. In this chapter, we present a hybrid
approach to conditional evidence fusion that leverages the avantages of both extensional
1
The rule of deduction (given ‘If P then Q’ infer ‘P therefore Q’).
181
Contents
and intensional approaches using subjective logic. Subjective logic theory provides opera-
tors for both abduction as well as deduction [100, 95, 140] thereby, making it possible to
as presented by Jøsang et al, in [100, 95, 140]. Given a conditional of the form ‘y ← x’ we
interpret the meaning as ‘IF x THEN y’. In traditional binary logic, this is seen in a truth-
functional manner following the truth table called ‘material implication’ such that any
antecedent x being true forces y to be evaluated as true as well. However, in practice there
are examples of false conditionals with false antecedent and true consequent. Therefore,
the more natural interpretation is that ‘The truth value of x is relevant to the truth value
of y’. In the sense of relevance connection between x and y, the interpretation should
also consider the case ‘y ← ¬x’ for completeness, so we can properly handle the case that
the antecedent x is false. Following classical probability theory [159], a conditional of the
form ‘y ← x’ can be interpreted in terms of probability calculus and can be expressed via
Note that, the notation p(y||x) is meant to indicate that the truth value of y is evaluated
with both the positive and the negative conditionals. This notation is only meaningful in
182
Contents
directly. In such cases, the required conditionals can be correctly derived by inverting the
available conditionals if we know the base rate p(y) (that is denoted as a(y) in subjective
a(y)p(x|y)
p(y|x) = (7.2)
a(y)p(x|y) + a(y)p(x|y)
opinion having full ignorance is called ‘vacuous’ opinion. It represents a measure of prior
Let the term ‘child frame’ denote a traditional state space of mutually disjoint states and
let the term ‘parent frame’ denote the evidence that was obtained. Bearing the notation
p(consequence|antecedent) in mind, if the ‘parent frame’ is about the antecedent and the
‘child frame’ is the consequent, then such inference is called ‘deduction’. Abductive reason-
ing is the case when the ‘parent frame’ is the consequent and ‘child frame’ is the antecedent
frames. However, in this section, we will focus on the binomial case only. Deduction on
Definition 42. (Deduction) [100] . Let ΘX = {x, x} and ΘY = {y, y} be two frames
Chapter 7
with arbitrary mutual dependency. Let wx = (bx , dx , ix , ax ), wy|x = (by|x , dy|x , iy|x , ay|x ) and
wy|x = (by|x , dy|x , iy|x , ay|x ) be an agent’s respective opinions about x being true, about y
being true given x is true and about y being true given x is false, then deductive conditional
183
Contents
where ⊚ denotes the general conditional deduction operator for subjective opinions and,
wy||x is defined by
by||x = bIy − ay K
bI = bx by|x + dx by|x + ix (by|x ax + by|x (1 − ax ))
d
y
y||x = dy − (1 − ay )K
I
where,
dIy = bx dy|x + dx dy|x + ix (dy|x ax + dy|x (1 − ax ))
iy||x = iIy + K
iI = bx i + dx i + ix (i ax + i (1 − ax ))
a y y|x y|x y|x y|x
y||x = ay
and K can be determined according to 3 different selection criteria detailed in [100].
Definition 43. (Abduction) [140] . Let ΘX = {x, x} and ΘY = {y, y} be two frames
(by|x , dy|x , iy|x , ay|x ) and wy|x = (by|x , dy|x , iy|x , ay|x ) be an agent’s respective opinions about
observed consequent y being true, vacous subjective opinion about the base rate of the hy-
pothesis x ,about y being true given x is true, and about y being true given x is false, then
abductive conditional opinion wx||y = (bx||y , dx||y , ix||y , ax||y ) about x being cause of observed
consequent y is expressed as :
where ⊚ denotes the general conditional abduction operator for subjective opinions, then
the inverted conditionals wx|y ,wx|y in the right hand side of the equation can be derived
wxvac · wy|x
wx|y =
wxvac ⊚ (wy|x , wy|x )
(7.6)
wxvac · ¬wy|x
wx|y = vac
wx ⊚ (¬wy|x , ¬wy|x )
thereby, we can calculate the ⊚ operation with the ⊚ operator and inverted conditionals.
Note that Equation 7.6 involves multiplication and division operators as well as de-
duction operator ⊚ (see Equation 7.4 and Table 3.2 in Definition 43). The multiplication
184
Contents
in subjective logic is also called ‘conjunction’ and the formal definition is shown in Defini-
tion 31. The inverse operation to multiplication is division. The quotient of opinions about
Definition 44. (Division /) [97] . Let ΘX and ΘY be two frames and let x and y
be propositions about state in ΘX and ΘY respectively. Let wx = (bx , dx , ix , ax ) and wy =
(by , dy , iy , ay ) be an agent’s opinions about x and y, then division opinion denoted as wx /wy
is wx∧y = (bx∧y , dx∧y , ix∧y , ax∧y ) such that :
x )ay (1−dy)
if ax < ay ,
if 0 < ax = ay ,
γ(1 − dx )
ay (bx + ax ix ) ax (1 − dx )
bx∧y = −
bx∧y =
(ay − ax )(by + ay iy ) (ay − ax )(1 − dy )
1 − dy
dx − dy
dx − dy dx∧y =
dx∧y = 1 − dy
1 − dy
ay (1 − dx )
(1 − γ)(1 − dx )
−
ay (bx + ax ix )
ix∧y =
ix∧y =
1 − dy
(ay − ax )(1 − dy ) (ay − ax )(by + ay iy )
a
ax∧y = ax /ay
x∧y = ax /ay
( )
ay (1 − ay ) (1 − dy )bx by
where, γ = − by + .
(ay − ax )(by + ay iy ) 1 − dx by + ay iy
The proposed hybrid conditional handling framework mainly relies on rule-based system
that enables logic programming proposed in Section 4.3.2. The rule-based system is ex-
Chapter 7
tended to allow representation of rules using subjective opinions and operators. The con-
ditional knowledge directly encoded as such rules are handled in extensional manner. To
handle conditionals in intensional manner, special types of predefined rules are introduced.
Such predefined rules drive the abduction and deduction operation in the subjective logic
framework in the presence of required prior opinions for the conditionals. We will first give
185
Contents
a brief overview how rules are expressed in logic programming. Thereafter, comes with
logic. In this section, we will briefly explain the framework again. In the proposed frame-
work, the CLIPS [1] rule engine was used as a basis to provide flexibility for defining
complex data structure as well as for providing a rule resolving mechanism. To extend
this system, a data structure ‘opinion(agent, proposition, b, d, i, a)’ was defined that can
be interpreted as a fact of arity 6 with the following terms, agent (opinion owner), propo-
the structure so that it can take arity n properties as well. Therefore, given a predicate p
the proposition can be described as ‘p(a1 , a2 , ..., an )’. In our system, therefore, each fact is
agent
represented as the form of ‘wp(a 1 ,a2 ,...,an )
’. Namely, rules are defined with the opinion and
as parameters were defined. Since in rule-based systems, actions can be executed in the
head part of the rule, the uncertainty assessment ‘ϕ : rules → opinion’ operation, defined
in Definition 16, can be defined in the head part of rules using subjective logic operators
and opinions shown in rule body. This aspect is depicted in Section 4.3.2 as follows :
Rule Body :
Due to the redundancy that arises when describing rules at the opinion structure level, we
186
Contents
wpacc(ac1 ,..,acn ) ← wpa11 (a11 ,..,a1n ) ⊛ ... ⊛ wpain (ai1 ,..,ain ) (4.2)
where ⊛ indicates one of subjective logic’s operators. More concretely, this implies checking
for the existence of each opinion that matches property constraints among the propositions
of opinions. In general, rules in the form of Equation 4.2 can also be seen as condition-
als. Thus, this way of representing rules and handling uncertainty can be considered as
extensional approach and in our framework such rules represent the extensional layer.
connection between x and y. To generalize this view, we will consider x and y as two facts
with arbitrary mutual dependency without restricting the influence direction. Then we
can set two conditionals ‘y ← x’ and ‘x ← y’. This results in 4 possible conditional
operations as shown in Table 7.1, namely, deduction and abduction for each conditionals.
In the sense of querying for the cases in Table 7.1, Case 1 can be interpreted as ‘given
an observational opinion x give me an opinion that this would make y happen’. Case 2
means ‘given an observational opinion y give me an opinion that it was caused by x’.
Case 3 is ‘given an observational opinion y give me an opinion that this would make
that it was caused by y’. As shown in Section 7.3, the calculation of those operations
requires incidental opinions (i.e. priors) also. According to Equation 7.4 and Equation 7.5
wx||y = wy ⊚(wy|x , wy|x , wxvac ) = wy ⊚ (wx|y , wx|y ) = wx||y by getting inverse opinions using Chapter 7
Equation 7.6. Thus, Case 2 and Case 3 are equal. Similarly, Case 1 and Case 4 are also
deduction on a proposition ‘x ← y’. Thus, given a single query shown in Table 7.1 we
have two choices of conducting this conditional depending on available priors. Therefore, a
decision making process, examining the availability of required priors in the rule and fact
187
Contents
Observation Priors
Query Operator
wx wy wx|y wx|y wy|x wy|x wxvac wyvac
√ √ √
Case 1 wy||x ⊚
y←x √ √ √ √
Case 2 wx||y ⊚
√ √ √
Case 3 wx||y ⊚
x←y √ √ √ √
Case 4 wy||x ⊚
Table 7.1: Possible Subjective Conditional Operations given x and y with Arbitrary Mutual
Dependency.
base, can be derived in form of a deterministic conjunctive logic program. Following this
saying if there exist three facts wx , wy|x and wy|x calculate wy||x . Following Equation 4.2,
the rule for Case 1 (i.e., Equation 7.7) will just be denoted as follows.
Other rules corresponding to the rest of the Cases 2 - 4 can be derived similarly. Note
that such rules are actually not about the context itself but about a mechanism how
the conditionals need to be handled and, therefore, those rules should be considered as
meta-rules. We consider those rules being part of the intensional layer. The CLIPS rule
engine provides prioritization of rules. By adjusting priorities of the meta-level rules and
other extensional rules, we can decide when the meta-level rules should be triggered. For
example, we could first apply the top priority group of extensional rules to infer occurrence
of some evidences that may be used for intensional layer. Then, we could intensionally set
meta-level rules to be the second priority group. Setting lowest priority group of extensional
rules, the inference result from intensional layer could be also used for extensional way of
inference. This way, the proposed approach can provide a combination of ‘extensional’ and
‘intensional’ approaches.
188
Contents
This section describes a case study for the application of forensic conditional reason-
ing using the proposed hybrid framework on metadata acquired using low level computer
vision analytics. Low level video analytic modules generate a rich set of metadata upon
which the proposed hybrid reasoning can be performed to detect more complex events.
This metadata captures the output of modules that detect ‘humans’ at a particular lo-
cation and the time, modules that detect humans passing, a pre-defined region etc. A
detailed description of these modules is outside the scope of this chapter. Figure 7.1 shows
frames from a video sequence captured from a visual surveillance system observing a typi-
cal airport scene. We also have access to information such as pre-annotated zones of stairs,
escalator, ATM machines, etc. Now assume we need to query this video sequence to detect
Chapter 7
all time instances when the escalator was non-functional. Please note, we do not have
any low-level computer vision module that can directly detect such an event. We wish to
formulate rules to detect this event based on circumstantial evidence of how the behavior
of the humans in the scene changes in response to the escalator not working.
In this case study we will show how we encode rules in the proposed hybrid reason-
ing framework to capture our knowledge of how we expect people behavior to change in
response to this event, and further show how we can use this knowledge to reason about
189
Contents
instances when the escalator is not-functional, given certain observations. Consider the
following rules :
At the same time we could also think about rather deductive knowledge as follows:
In this scenario, knowledge about when people use the escalator or the stairs is essential.
Therefore, based on the output of low level vision modules, we need to derive rules to
reason about the tendency of ‘people use escalator’ or ‘people use stairs’. Consider the
following rules:
190
Contents
These rules capture our assumption that humans tend to always be in a state of ‘action’
when interacting with certain structures within a building. Therefore, when a low level
vision module detects a human, at a particular time instant, on a structure such as stairs
Once such an evidence is built, we can assess the plausibility (opinion in Subjective Logic)
about escalator’s working or not using either Case 1 (abductive rules) or Case 2 (deductive
This section deals with rule modeling for the scenario, based on the system setting de-
scribed in Section 7.5.1. In both conditional knowledge models shown in Section 7.5.1,
the observation about people use escalator or stairs serves as basis for the assessment.
Therefore, we first start with observational knowledge model described in Section 7.5.1.
Assuming opinions about observational metadata from human detection and annotation,
HumanDetector
we will use opinions such as whuman(H Annotation
, wescalator(E Annotation
i ,LocH ,Tj )i 1 ,LocE 1
,Tj ) , wstairs(S1 ,LocS 1
,Tj ) ,
etc (where for example, the first notation says, there is an opinion from HumanDetector
about a proposition human (i.e., a human exists, more concretely) with id Hi and local-
a pre-annotated zone Zonek in the scene, low level modules also generate an opinion of
GeometryAgent
the form wwithin(H i ,Zonek ,Tj )
. We combine these opinions with the conjunction operator ∧
Chapter 7
wa_human_uses_escalator(Hi ,Tj ) ←
(7.9)
GeometryAgent
HumanDetector
(whuman(Hi ,LocH ,Tj )
∧ wescalator(E
Annotation
1 ,LocE ,Tj ) ∧ wwithin(Hi ,E1 ,Tj ) )
i 1
wa_human_uses_stairs(Hi ,Tj ) ←
(7.10)
GeometryRule
HumanDetector
(whuman(H i ,LocH ,Tj )
∧ wstairs(S
Annotation
1 ,LocS ,Tj ) ∧ wwithin(Hi ,S1 ,Tj ) )
i 1
191
Contents
While above Rule 7.9 and Rule 7.10 will fire for every single human detection, we are
interested in accumulating all the instances of evidences from these rules, to get a single
opinion about ‘people_use_escalator’. An opinion from Rule 7.9 or Rule 7.10 is too weak
‘reputation (discount)’ operator ⊗ [99] (see Definition 41) to every single detections as
follows.
byHi
wpeople_use_escalator(Tj)
←
a_human_uses_escalator
(wa_human_uses_escalator(Hi ,Tj ) ⊗ wpeople_use_escalator )
(7.11)
byHi
wpeople_use_stairs(Tj)
←
a_human_uses_stairs
(wa_human_uses_stairs(Hi ,Tj ) ⊗ wpeople_use_stairs )
Then, we need to fuse each of decisions made from above Rule 7.11. Using consensus
operator ⊕ [93, 94] in subjective logic (see Definition 30 Section 5.6.3), we will combine
both cases of people use escalator or stairs into one observational opinion. The encoded
rule is as follows.
wpeople_use_escalator(Tj ) ←
⊕ n (7.12)
byHi byHi
(⊕ni=1 wpeople_use_escalator(Tj)
) (⊕i=1 ¬wa_people_use_stairs(Tj)
)
Rule 7.12 cumulates every single human detection’s contribution to the opinion about ‘a
human uses escalator’. In the similar way, it also cumulates every single human detec-
tion’s contribution to the opinion on ‘human uses stairs’. Both cumulated opinions are
again fused by consensus operator. In the context that we fuse both positive and negative
evidences, it resembles the subjective logic based ‘default inference’ (see Definition 35).
Furthermore, this is also similar to the binomial observation that deals with number of
positive and negative evidences (see Section 3.1.3), but with uncertainty consideration.
For the simplicity of illustrative explanation, assume that we had perfect human detector
and perfect geometry agent, so every results from Rule 7.9 and Rule 7.10 were definite true.
192
Contents
a_human_uses_escalator a_human_uses_stairs
Setting both the discount opinions as wpeople_use_escalator = wpeople_use_stairs =
byHi byHi
(0.5, 0, 0.5), every single opinion of wpeople_use_escalator(Tj)
and wpeople_use_stairs(Tj)
also
becomes (0.5, 0, 0.5). Then, based on this assumption, Figure 7.2 shows examples of cal-
a given time Tj .
As explained in Section 7.4.1 these rules belong to ‘extensional layer’. Now, to model
the conditional knowledge models from Section 7.5.1 that work at the ‘intensional layer’,
and y as people_use_escalator, we could assign opinions for the items shown below.
•wescalator_working|people_use_escalator •wescalator_working|¬people_use_escalator
Once Rule 7.9 ∼ Rule 7.12 are triggered, based on available information among above
priors and according to Table 7.1, appropriate conditional reasoning rules are dynamically
triggered. Although, there exist formal ways of extracting priors from data, we will for this
case study, use pre-defined subjective prior opinions for the convenience of our discussion
and considering the scope of this chapter (note that, however, in fact the use of such a
193
Contents
subjective opinion also makes sense in the view of subjectivism in Bayesian, that uses
Example 8. (Is the Escalator Working?) . In the case of defining usual tendency
(base rate) of escalator’s working to 90%, opinion about people would use escalator when an
escalator is working to (0.93,0.01,0.06) and opinion about people would still use escalator
194
Contents
w ←
escalator_working(E1 ,Tj )||people_use_escalator(Tj )
Figure 7.3 depicts graphical model of introduced conditionals for the case study and Fig-
ure 7.4 shows an illustrative example of Rule 7.13 in the case of observational opinions
from Rule 7.9 ∼ Rule 7.12 were as depicted in Figure 7.2. In the case we observed 1 person
using escalator and 1 person using stairs, by the base rate and prior, we get more belief on
escalator working but with high uncertainty. As we observe more people using stairs, we
As proof of concept, a demonstrator has been built. CLIPS rule-engine was used to imple-
ment subjective logic extended logic programming for the proposed reasoning framework.
Eclipse platform was used to integrate other modules such as opinion visualization, video
parser, metadata parser and manipulation UI, etc. At this stage of case study, we used 300
seconds of airport video surveillance data. Then we manually annotated human detection
metadata such as bounding box and time information per seconds as shown in Figure 7.1
with a background annotation (e.g. escalator, stairs, etc). We first set the necessary pri-
ors as shown in Figure 7.4. To consider each of single human detection based judgement
as a reasonably weak and uncertain evidence, we used strong discount opinion factor as Chapter 7
a_human_uses_escalator a_human_uses_stairs
wpeople_use_escalator = wpeople_use_stairs = (0.3, 0.05, 0.65). For more intuitive case
study, we first considered a real-time analysis setup. Meaning that we loaded Rule 7.9 -
7.13 and injected low level metadata along the video playing in synchronized manner. To
see how belief changes over time, we additionally introduced a rule as follows.
195
Contents
Figure 7.5: Screen Captures demonstrating Belief Revision on the Proposition ‘the escala-
tor is not working’.
196
Contents
Accumulated_Over(T ,T )
1 n
wpeople_use_escalator(Tj)
←
(7.14)
Accumulated_Over(T ,T
1 n−1 )⊕
wpeople_use_escalator(Tn−1 )
(wa_people_use_stairs(Tn ) )
Then we kept tracking the change of the derived opinion about the proposition ‘escala-
tor is not working’. When the system gets opinions about human detection, rules were
automatically triggered. Figure 7.5 shows the change of opinion on ‘escalator is not work-
ing’ along a timely sequential multiple observations. For example, the 4th observation in
Figure 7.5 shows a scene that many people were using stairs while nobody was using the
escalator. In this case, the system correctly computed a strong opinion that the escalator
is not working. While, 5th observation shows an example of retracting current opinion
terms of ‘belief revision’ over time. The demo on this case study shows that it well simulate
human intuition.
In this section, we further explorer the proposed approach with a virtual scenario.
Let us consider a conceptual scenario that a security personnel wants to find a scene of
parking a vehicle, that seems to be done by a ‘novice driver’. Let us also assume a virtual
system setting such that: parking slots are pre-annotated with a polygon, vision analytics
can detect moving and stopped vehicles with motion vector and geometric localization Chapter 7
information (i.e., coordinate, width and height). Given the assumption, Figure 7.6 shows
some of sequential video footages 2 . In Figure 7.6, parking slots are annotated with blue
polygons and vehicles with no motion vectors are labelled with pink ovals. Vehicles with
motion vectors are marked with red boxes and red arrows with its vehicle ‘id’. Humans
2
The video footages are inspired and intensionally synthesized based on the original YouTube video on
http://www.youtube.com/watch?v=_rHY1qKLLws&feature=related.
197
Contents
can easily infer lots of implied high level semantics of Figure 7.6 as follows.
‘It seems that the car C1 was trying to park to an empty parking slot EP S1 several times.
In the meanwhile the car was trying to park, another car C2 appeared. The newly appeared
car C2 seems to be blocked due to the car C1, that is still trying to park (see t9 ∼ t12 in
Figure 7.6). Because the car C1 was very sluggish at parking, it seems that the driver of the
car C1 was a novice driver. If the driver of C1 was a novice driver, there is a possibility
that the car C1 could have scratched other cars. If such case was really occurred, the car
C2 would be a good witness to get more information on the scene and on the driver of C1,
etc.’
In this case study, we will show how we encode some of above knowledge as rules in the
198
Contents
At the same time, we could also think about rather deductive knowledge as follows:
• Rule 3 - "if he/she is not good at parking, usually a novice driver is driving" or
• Rule 4 - "if he/she is reasonably good at parking, usually not a novice driver"
In this scenario, the knowledge about ‘whether a parking was good or not’ is essential.
Therefore, based on the output of low level vision modules, we need to derive rules to
get opinions on ‘good at parking’ or ‘not good at parking’. Consider the following rules:
• Rule 5 - "if a car comes in to, and goes out of an empty parking slot multiple times,
‘¬ good parking ← areas of car and empty slot are overlapped multiple times’ or
‘good parking ← areas of car and empty slot are overlapped only f ew times’
3
Note that, in this scenario, by assumption, we do not have any low-level computer vision module that
can directly detect the high level semantics such as ‘novice driver’, etc.
199
Contents
The implication of Rule 5 and 6 is similar to the one of binomial observations in the context
that the Rule 5 and 6 also observe multiple times of positive evidences (see Section 3.1.3
for more details on binomial observation). However, in our context, Rule 5 and 6 only
matter positive evidences (i.e., occurrence of ‘geometric overlap between the car and the
empty parking slot of interest’). It can be also stated as ‘the more times the car tries to
park, the higher is the likelihood that it was driven by a novice driver’ which is linguistically
similar to the ‘vague rules’ discussed in the previous Chapter 6. However, while Rule 5
and 6 deal with the number of occurrence of the event (defined in rule head) itself, the
vague rules discussed in Chapter 6 deal with an attribute implied in the condition (rule
body) to occur an event (defined in rule head) 4 (namely, vague rules in Chapter 6 concern
about intrinsic parameters within a ‘predicate’ based proposition5 ). Therefore, the Rule 5
and Rule 6 should be designed as an increasing function of the number of ‘parking trial’.
Namely, the more observational instances of evidence on ‘parking trial’ should entail the
Based on the discussion of previous Section 7.6.1, this section deals with actual rule mod-
eling for the given scenario. In both conditional knowledge models shown in Section 7.6.1,
the observation about the goodness of parking serves as basis for the assessment. Therefore,
we first start with observational knowledge model described in Section 7.6.1. Assuming
opinions about observational metadata from vehicle detection and annotation, we will use
V ehicleDetector
opinions such as wvehicle(V Annotation
i ,LocV ,M VV
i 1
,Tj ) , wparking_slot(P Sk ,LocP S k
,Tj ) , etc (where for exam-
ple, the first notation says, there is an opinion from ‘VehicleDetector’ about a proposition
‘vehicle’ (i.e., a vehicle exists, more concretely) with id V1 , localization information LocV1
4
In Chapter 6, ‘direction’ and ‘distance’ are the concerned attributes.
5
See Section 3.2.2 for detailed explanation about the difference between normal propositional logic and
first-order predicate logic. Also note that, the rule engine used in our framework is based on the
first-order predicate logic (see Section 4.3.2).
200
Contents
within the parking slot, examining whether the motion vector of the car is Null. Therefore,
byVi
woccupied_parking_slot(V i ,P Sk ,LocP S ,Tj ) ←
k
GeometryAgent
V ehicleDetector
(wvehicle(Vi ,LocV ,N U LL,Tj )
∧ wparking_slot(P
Annotation
Sk ,LocP S ,T ) ∧ wwithin(Vi ,P Sk ,Tj )
) (7.15)
i k j
byVi byVi
wempty_parking_slot(P Sk ,LocP S ,Tj ) ← ¬woccupied_parking_slot(_,P Sk ,LocP S ,Tj )
k k
As indicated by the agent markup ‘byVi ’, Rule 7.15 derives an opinion about the proposition
a vehicle ‘Vi ’ and a parking slot ‘P Sk ’. By default, however, every parking slots have
chances of being empty, even if there are no clues about occupancy. Consider following
truth assignment where, DT indicates ‘default true’, where any value reasonably close to
‘true’ but rather ‘uncertain’ can be a candidate of DT as long as we fix the value throughout
reasoning pipeline (see Definition 16 and Definition 35). In this case, following examples
In this setup, following the ‘default reasoning’ scheme introduced in Chapter 5 (see Defi-
Now, we introduce a rule to know whether a car is trying to park. Consider a rule as
follows.
byV
_parking(Vi ,P Sk ,Tj ) ←
i,j
wtrial_of
byDef aultInf erence GeometryAgent
V ehicleDetector
(wvehicle(Vi ,LocV ,N U LL,Tj )
∧ wempty_parking_slot(P Sk ,LocP S ,Tj ) ∧ woverlap(V i ,P Sk ,Tj )
)
i k
(7.18)
201
Contents
While above Rule 7.18 will fire for every single parking trial, we are interested in accu-
mulating all the instances of Rule 7.18, to get a single opinion about ‘good_at_parking’.
Therefore, the accumulation should form a decreasing opinion function6 in terms of number
of observations. To achieve this, we first apply the ‘reputation (discount)’ operator ⊗ [99]
(see Definition 41) to every single parking trial detection as follows, because a single in-
byV
i,j
wgood_at_parking(Vi ,P Sk ,Tj )
←
(7.19)
byV trial_of _parking
_parking(Vi ,P Sk ,Tj ) ⊗ wgood_at_parking )
i,j
(wtrial_of
Then, we need to fuse each of decisions made from above Rule 7.19. By default, however,
it would be expected to be a normal parking, and the number of parking trials would form
a negative evidence against good parking. Therefore, using consensus operator ⊕ [93, 94]
in subjective logic (see Definition 30 Section 5.6.3), the encoded rule becomes as follows.
⊕ byV
wgood_at_parking(Vi ,P Sk ,Tj ) ← DT ¬(⊕nj=1 wgood_at_parking(V
i,j
i ,P Sk ,Tj )
) (7.20)
to the opinion on ‘good at parking’. By the nature of consensus operator ⊕ and ¬, the more
instances from Rule 7.18 and Rule 7.19 (i.e., ‘trial_of_parking’) we detect, the weaker belief
on ‘good_at_parking’ we get. For the illustrative explanation of this aspect, assume that we
had perfect vehicle detector and perfect geometry agent, so every results from Rule 7.19
trial_of _parking
were definite true. Setting the discount opinion as wgood_at_parking = (0.33, 0, 0.67)7 ,
i,j byV
every single opinion of wgood_at_parking(Vi ,P Sk ,Tj )
also becomes (0.33, 0, 0.67). Then, based
the by Rule 7.15 ∼ Rule 7.20, given different number of parking trials of Vi . In Figure 7.7,
as more instances of parking trials are considered, Rule 7.20 set stronger disbelief on
6
It becomes increasing opinion function in case we consider ‘NOT good at parking’.
7
The implication of this assignment is to consider at least three times of parking trial to decide possibility
of being a bad parking. Therefore, by intuition, we assigned 1/3 belief of full belief.
202
Contents
‘good_at_parking’ proposition.
Thus far, we have reviewed rules belong to ‘extensional layer’. Now to address the
conditional knowledge models from Section 7.6.1 that work at the ‘intensional layer’, we
need to collect certain prior opinions. In our case, by setting x as novice_driver and y as
•wnovice_driver|good_at_parking •wnovice_driver|¬good_at_parking
•wgood_at_parking|novice_driver •wgood_at_parking|¬novice_driver
•wnovice_driver
vac •wgood_at_parking
vac
Chapter 7
Once Rule 7.15 ∼ Rule 7.20 are triggered, given available priors, one of following meta-level
w ←
novice_driver(Vi )||good_at_parking(Vi )
203
Contents
wgood_at_parking(Vi )||novice_driver(Vi ) ←
wnovice_driver(Vi )|¬good_at_parking(Vi ) )
Note that, Rule 7.21 and Rule 7.22 use different priors, but returns an opinion on the same
In this section, we show conceptual reasoning example given the scenario, rules and as-
Example 9. (Novice Driver) . Assume that we have detected a vehicle with id C1 and a
parking slot EP S1 as depicted in Figure 7.6. Let us also assume that we have observational
that 10% of drivers are novice driver in general. Consider the case we have intuition based
•wnovice_driver
vac = (0.0, 0.0, 1.0, 0.10) (i.e., 10%)
then, according to Table 7.1, Rule 7.21 will be triggered. Given the assumptions and the
priors above, Figure 7.8 shows an illustrative example of applyging Rule 7.21.
In the case there was only one observational instance on ‘good_at_parking’, the com-
putational result yields more ‘disbelief ’ than ‘belief ’ about the proposition ‘novice_driver’.
opinions, it smoothly changes its opinion toward stronger belief about the proposition
‘novice_driver’. This shows that the reasoning results coincide with human intuition.
8
Note that, in fact, such subjective opinions (based on human intuition) can be regarded as subjective
Bayesian priors as discussed in Section 3.1.3. Refer to [33] for details of Bayesian statistics and its
comparison with traditional frequentist methods.
204
Contents
Example 10. (Illegal Parking) . Consider the same situation as the one of Example 9.
Assume that we have detected a new vehicle with id C2. C2 slowly entered in the scene and
suddenly stopped after a while, as depicted in the footages whthin the period (t9 ∼ t12)
in Figure 7.6. Assume a set of rules about determining ‘illegal parking’ and the opinion
assignments to the rules as follows.
byVi
ϕRule [¬w¬illegal_parking(Vi ,Tj ) ← wblocked_vehicle(V i ,Tj )
] = (0.9, 0, 0.1)
GeometryAgent
wblocked_vehicle(Vi ,Tj ) ← wnovice_driver(Vi ) ∧ wmotion_vector_bef ore_stop_heading(Vi ,V j)
(7.23)
byVi
where, woccupied_parking_slot(Vi ,P Sk ,LocP S ,Tj ) is the opinion determined by the Rule 7.15
k
and wnovice_driver(Vi ) is the opinion after Rule 7.21 and Rule 7.22 are triggered in the
‘intensional layer’.
205
Contents
Consider opinion assignments for the case that C1 has tried to park 4 times. For
this example, we will take the corresponding opinion wnovice_driver from the Figure 7.8 of
Example 9. Also assume that our geometry agent is robust and reliable.
#of inst.=4
ϕRuleEval [wnovice_driver(C1) ] = (0.32, 0.22, 0.46)
GeometryAgent
ϕf act [wmotion_vector_bef ore_stop_heading(C2,C1) ] = T = (1, 0, 0)
V ehicleDetector
ϕf act [wvehicle(C2,Loc C2 ,N U LL,Tj )
] = T = (1, 0, 0)
Given this setting, the inference on ‘C2 parked illegally’ can be conducted as follows.
clsldi (ϕ)(willegal_parking(C2) )
= [U ⊔ ((1, 0, 0) · DT1 )] ⊕ ¬[U ⊔ (((0.32, 0.22, 0.46) · (1, 0, 0)) · DT2 )]
= [U ⊔ (1, 0, 0) · (0.2, 0, 0.8)] ⊕ ¬[U ⊔ ((0.47, 0.22, 0.31) · (0.95, 0, 0.05))]
= [U ⊔ (0.47, 0, 0.53)] ⊕ [U ⊔ ¬(0.55, 0.22, 0.23)]
= (0.47, 0, 0.53) ⊕ (0.22, 0.55, 0.23)
= (0.35, 0.46, 0.19)
The entailed opinion (0.35,0.46,0.19) can be linguistically interpreted as ‘it may not be
an illegal parking, but with some amount of uncertainty’. In the same way, the other
stopped vehicles labelled with pink ovals can be regarded as ‘illegal parking’ when they
are not parked within a predefined parking slot (see Figure 7.6). This example shows that
the inference result from the ‘intensional layer’ can be again used for another semantic
premises can be factorized and handled in a Bayesian sense by using subjective logic’s
deduction and abduction operators. The main advantage of the proposed hybrid approach
is that it offers more choices to handle the trade-offs between expressive power, semantic
clarity and computational efficiency. However, there are still several open issues such as
206
Contents
how to extend this concept to multinomial frame based event, how to automatically get the
inference under uncertainty, for high level semantic analysis of visual surveillance data. For
uncertainty handling, subjective logic was adopted on top of rule-based framework. The
main advantage of the proposed hybrid approach is that it offers more choices to handle
the trade-offs between expressive power, semantic clarity and computational efficiency. For
highly varying and complex compositional events in visual surveillance data, we could ben-
efit from the expressive power and modularity of the extensional approach, while leveraging
There are, however, still several open issues such as how to extend this concept to
multinomial frame based event, how to automatically get the prior opinions, etc. Although
we still need to extend this concept to large scale data. We advocate that this work showed
the potential of the proposed approach. One of interesting properties of the system is
that, unlike traditional probability based conditional reasoning, this approach allows for
representing lack of information about a proposition. We could assign priors with lack
of information, and observations can also be represented with any degree of ignorance,
therefore we believe this better reflects human intuition and real world situations. Another
beneficial property is the flexibility of defining rules by taking a layered approach to leverage
207
208
8 Logical Abduction under Uncertainty
Thus far, our discussions and reasoning examples have dealt reasoning scenarios that can be
be reasoned about but just by considering set of observations of their interest. Such a
reasoning is called logical abduction and this chapter represents our approach to achieve
8.1 Introduction
hypotheses for a given set of collected observations for forensic visual surveillance. As video
analytic power exploited in visual surveillance is getting matured, the more automatically
generated intermediate semantic metadata became available. In the sense of forensic reuse
of such data, the majority of approaches have been focused on specific semantic query
based scene analysis. However, in reality, there are often cases in which it is more natural
to reason about the most probable semantic explanation of a scene given a collection
reasoning pipeline that combines abductive logic programming together with backward
Chapter 8
and forward chaining based deductive logic programming. To rate derived hypotheses, we
1
note that, while the subjective logic ‘abduction’ operator discussed in Chapter 7 deals the ‘bidirectional
interpretation’ of a conditional rule, ‘logical abduction’ deals a diagnostic mechanism to find most
probable hypothesis given set of observations
209
Contents
apply subjective logic. We present a conceptual case study in a distributed camera based
scenario. The case study shows the potential and feasibility of the proposed approach for
As discussed in Section 1.1, recent advances in computer vision technology have made
surveillance. When it comes to forensic reuse of such analytics, however, it is still beyond
the sufficient intelligence due to the variety of possible semantics and complex plots implied
requires intelligent reuse of low level vision analytic results in a context sensitive manner.
To address this aspect, there has been some work on the use of declarative logic
formalism to represent and reason about high-level contextual semantic knowledge which is
appear in literature. After the review and comparison, in Chapter 4, we have also proposed
an ‘extensional’ approach that uses logic programming with subjective logic theory.
In such approaches, reasoning is triggered with a specific high level semantic query
to decide the existence of metadata patterns that semantically satisfy the given query.
However, in reality, there are often cases require to ask about the most probable semantic
explanation of a scene based on specific evidential semantic observations. That is, when
the former approach is applied to semantic retrieval, the security personnel is required to
know what semantics to ask. While the later would, ideally, let users collect evidences
of their interest and do simplify queries into a single ‘why?’ or ‘what?’ type of query.
Enabling such type of reasoning would be advantageous especially in the situation that
a scene are possible. In general, this type of diagnostic reasoning is known as abduction
[131]. In this chapter, we present a layered pipeline of reasoning flow that adopts abductive
210
Contents
logic programming together with backward and forward chaining based deductive logic
programming and combine subjective logic [92, 93] for the rating of derived diagnostic
hypotheses. We then present a conceptual case study from a distributed camera based
scenario to show the potential and feasibility of the proposed approach in the sense of
diagnostic forensic.
To achieve better expressive power and flexibility on context modelling and reasoning of vi-
sual surveillance data, there has been some work on the use of declarative logic formalisms
and on the use of different uncertainty handling formalisms. As we have already reviewed
in Section 2.4.3, Akdemir et al. [13] used an ontology for human activity recognition, but
without uncertainty handling. Shet et al. [152, 153] proposed a system that adopts Prolog
based logic programming extended with the bilattice framework [69] for default logic [143]
based situation reasoning and human identity maintanance. Jianbing et al. [110] used
rule-based reasoning with Dempster Shafer’s belief theory [149] for a bus surveillance sce-
nario. Anderson et al. [17] used fuzzy logic [177] to model human activity for video based
eldercare. Based on our previous work, [78, 81, 80, 82, 83, 79], in this dissertation, we
proposed the use of logic programming and subjective logic [92, 93] to encode contextual
ence and default reasoning for visual surveillance scenarios. In the sense of reasoning, such
logic framework based approaches require a specific query to reason about, and yield an
However, in the sense of forensic use of visual surveillance data, rather diagnostic
reasoning so called ‘logical abduction’ is also required. In such diagnostic reasoning, the
Chapter 8
initial input to the reasoning system would be desired to be a selected set of interested
event, then the query can be converged into a single ‘why?’ or ‘what?’ type of query to
get explanations or hypotheses to the best of given knowledge base. There has been also
211
Contents
some work on the use of abduction. Espinosa et al. [132] showed description logic [2]
based abduction example getting explanations about an image given ontology based image
annotation. Ferrin et al. [59] also used description logic for visual surveillance to show that
abduction can be used for realtime ‘car theft’ event detection. While both work showed the
potential of abduction in interpreting media contents, the use of description logic makes
can also conduct other types of reasoning such as default reasoning. In contrast to the
previous work, we take more pragmatic approach based on logic programming including
abductive logic programming and rate results with subjective logic with more focus on
for a given set of observations. Peirce [131] showed that it can be seen as inverse modus
represents observations, ∆ represents the explanations to be computed and the symbol ‘|=’
represents logical entailment. During the 90s this aspect has been solidly studied in the field
of AI in the context of logic programming and came up with abductive logic programming
logic program P , a set of ground abducible atoms A and a set of classical logic formulas
IC, called the integrity constraints, such that no p ∈ A occurs in the head of a rule of P .
The logic program P also consists of rules R and literals L, namely, P = (R, L). Then
for a given query observation Q (note that, in abductive logic programming, Q is not a set
of observations but rather a single observational query), an abductive explanation for the
212
Contents
consistent. This mean that the derived explanation ∆ and P should be able to entail Q.
At the same time, should not violate the constraints IC. In the sense of Peirce’s definition
of abduction, (P, A, IC) corresponds to Σ and Q is the Γ and ∆ is the output. Following
Example 11. (Abduction) . Assume an abductive logic program (P, A, IC), where
given the constraint IC and the fact the_sun_was_shining in P . (note that, in normal
logic programming the output is truth value about the query). There has been some work
for showing whether or not a given set of facts are valid according to currently known rule
SLDNFA [51] is one of well known algorithms to fulfill abduction itself. ACLP [101],
CIFF, SCIFF [6] and Asystem [125] are some of known implementations of abductive logic
Chapter 8
framework. Besides, ABDUAL and ProLogICA., etc., appear in literature but the project
sites are no longer appear. Indeed, it seems that no prominent benchmark or comparison of
those systems appear. Although some are based on Prolog, some of them are also based on
213
Contents
other languages such as ECliPSe language [57]. In this dissertation, considering availability
of executable binary, simplicity on system integration issues, we have used ACLP [101].
The objective of the presented framework is to enable abductive semantic reasoning and
its quantitative evaluation given multiple observational evidence input. For the core part
of the logical abduction, abductive logic programming plays an important role. As shown
in Section 8.4, however, abductive logic programming also requires a specific single query
to be explained and this is far from our intention. Although we could collect a set of
interested observations and put them as part of facts (literals L) in P of (P, A, IC), still
the reasoning should be triggered by a single query. To address this aspect, we take
and evaluation of derived hypotheses. To process the initial multiple observational input,
we adopt a forward chaining based rule engine to derive a possible set of semantics to be
asked. Unlike backward chaining based rule engine such as prolog that only tracks back
patterns that satisfy the given single query, a forward chaining based rule engine derives
all deductive logical conclusions given input literals and rule set at one execution. For
this part, CLIPS [1] rule engine was used. Once we get the list of possible semantics
to be asked and explained, for each of them we conduct the abduction and get a list of
logic programming framework. For the prototyping purpose, ACLP [101] implementation
of abductive logic programming was used due to its relatively simple structure among
other frameworks such as ASystem [125]. For evaluation of the derived hypotheses we
adopt subjective logic extended deductive rule based framework. Namely, for each of the
potential explanations, we conduct backward chaining based reasoning with subjective logic
based uncertainty evaluation. This layered pipeline is depicted in Figure 8.1. For input
video sources with analytic results, we first select a set of interested evidences to form Γ,
214
Contents
the set of observations which is then sent to the forward chaining based rule engine to
reason about possible semantics to be asked. For each of the semantics to be asked, we
form a single query Q which is sent to the abductive logic programming engine. Once we
get explanation set ∆, this is evaluated with the subjective logic based backward chaining
This section deals with an illustrative case study on the application of forensic abductive
reasoning in a typical distributed camera based scenario. We will assume metadata ac-
quired from low level computer vision analytics such as human detection, vehicle detection
and object detection, etc. As explained in Section 8.4 and Section 8.5, in our setup, such
metadata will be considered as logical facts (literals L, in the sense we care both positive
and negative facts) and will be put into logic program P together with context rules R
(note that, P = (R, L)). Assuming a security personnel examining a video scene for certain
forensic reason, the query for the abductive reasoning will also be assumed to be triggered
by Γ, the set of collected evidential metadata on his interest. Given the Γ, we first use
forward chaining based deductive logic programming to derive possible semantic queries
Qi to be explained and will be sent to the core abductive logic programming part to get
Chapter 8
the set of explanations ∆ij . Then finally, we will evaluate the resulting ∆ij with subjective
215
Contents
Figure 8.2: The Setup for Case Study Scenario with 3 Cameras Deployed.
Example 12. Scenario (Why the person suddenly went outside ?) . Assume a
typical distributed visual surveillance with three cameras deployed as shown in Figure 8.2.
Cam1 monitors the lobby of the building, Cam2 and Cam3 monitor front of the gate in
different outside views. Suppose a security personnel was watching a scene of Cam1 and
got interested in why the person who was sitting at the front desk suddenly stood up and
went outside the building. From the scene of Cam1, he collected metadata such as the
person’s standing up and heading to the predefined gate zone as depicted with the box
Given the Γ, suppose the system had the following deductive logic program segment.
216
Contents
The use of the forward chaining rule engine with the given Γ will then assert new semantic
derivations Qi as follows.
Q1 = go_outside(pid1)
reasoning framework. Suppose also that, at the same time of the event in Cam1, there
were a vehicle parked in front of the gate and also a chair beside the vehicle. When the
vehicle leaves, it hit the chair and the chair fell down on the road. These are captured
in Cam3 in Figure 8.2 and the analytic metadata of the scene is stored in form of literals
L. Consider an abductive logic program (P, A, IC) containing context rules R, integrity
The rule set R means that, a person may go outside to do something outside or because
the person saw some event. Taking a vehicle or buying something can be seen as doing
217
Contents
something in case that there was a vehicle to take or a mart to go to buy something.
Similarly, if there was an accident in which a vehicle and an object are involved and the
person is trying to manage the accident somehow, it implies that the person noticed the
event. This is also same in the case of an accident in which a vehicle and a person were
involved. In these cases, both a vehicle and an object or both a vehicle and a person
should be existing at the same time. As a constraint, the integrity constraint rule IC
means that some of those event should happen exclusively. Given the knowledge, the
candidate semantic predicates we may want to derive by the execution of abduction are
designated in the set of abducibles A. Now, with the given above abductive logic program
and an input query Q1 = go_outside(pid1), the abductive logic procedure will derive
explanations as follows.
In above scenario, we get two hypotheses sets for a given query Q1 . In the next step, we
examine each of the elements in ∆11 and ∆12 by the use of the subjective logic extended
logic programming as explained in Section 4.3.2. Consider following logical rule segments.
wtake_a_veh(P1 ,V1 ) ←
Human_Detector V ehicle_Detector
whuman(H1 ,T1 ,X1 ,Y1 ) ∧ wvehicle(V1 ,T1 ,X2 ,Y2 ) ∧ (8.1)
Geometry_M anager
woverwlap(X1 ,Y1 ,X2 ,Y2 )
wmng_veh_obj_acc(P1 ,V1 ) ←
Human_Detector V ehicle_Detector
whuman(H1 ,T1 ,X1 ,Y1 ) ∧ wvehicle(V1 ,T1 ,X2 ,Y2 ) ∧
(8.2)
Object_Detector Geometry_M anager
wobject(O1 ,T1 ,X3 ,Y3 ) ∧ woverwrap(X2 ,Y2 ,X3 ,Y3 ) ∧
Geometry_M anager
wnear(X1 ,Y1 ,X3 ,Y3 )
Rule 8.1 corresponds to the abducible take_a_veh(pid1) in ∆11 and Rule 8.2 corresponds
218
Contents
rules and metadata in form of literals L, both rules can be satisfied. Another aspect to
consider is the negative conditions required not to satisfy in ∆. In above cases, ∆11 requires
3 exclusive abducibles not to be satisfied and ∆12 requires 2 exclusive abducibles not to be
satisfied. Indeed, due to the uncertainty implied in metadata, rest of the abducibles with
negation could also come up with certain amount of uncertainty that are also represented
as subjective opinions. In this case, we will take a subjective logic’s complement operation
on the derived opinions. In case we can’t even prove an abducible predicate in ∆ given
logic program P , we will simply assign full ignorance opinion. Considering the positive
abducibles and negative abducibles to be satisfied, it can be seen as default reasoning [143]
and, therefore, we will conduct default reasoning using subjective logic [82, 83] introduced
in Chapter 5. In this work, we slightly modify the reasoning scheme shown in Definition 35
∧ ⊕ ∧
cl(ϕ)(∆) = [ w(p) ] ¬[ w(p) ] (8.3)
p∈∆ ¬p∈∆
operators respectively.
Now, consider the case that uncertainty assignment is as follows, that is, having the
same amount of opinion on each positive evidences in ∆11 and ∆12 . Namely,
Assume that, the system failed to prove mng_veh_prs_acc (pid1) and come up with
Chapter 8
buy_smt(pid1) with relatively high confidence that it was not happened, as opinion as-
signment follows:
219
Contents
Given this set up, applying Definition 45, the final evaluation results of each ∆i is as fol-
lows:
∆op
11 = cl(ϕ)(∆11 ) = (0.6, 0.3, 0.1)⊕
¬[(0, 0, 1) ∧ (0.6, 0.3, 0.1) ∧ (0.1, 0.8, 0.1)]
∆op
12 = cl(ϕ)(∆12 ) = (0.6, 0.3, 0.1)⊕
¬[(0.6, 0.3, 0.1) ∧ (0.1, 0.8, 0.1)]
In above example, the final result rated that the hypothesis ∆12 is a bit more plausible
than ∆11 (∵ ∆12 has more belief and less uncertainty than ∆11 , the expectation value Exp
of ∆12 is therefore, greater also). However, the reason the opinion gap between these two
∆1 ’s are not bigger is due to the limited set of rules and metadata we assumed. Therefore,
thorough domain knowledge engineering and rule derivation would be critical on applying
abduction based reasoning. It mainly consists of tow layers. 1) a layer that reason about
queries to be asked and 2) a layer that uses abductive logic programming to draw potential
hypotheses given set of observations. Our demonstration showed the feasibility of using
the proposed approach on diagnostic queries. For practical use of this approach, however,
220
Contents
surveillance. For logical abduction, abductive logic programming framework was adopted
together with forward chaining based rule engine. The resulting hypotheses are evaluated
by a rule based system that is extended with subjective logic operators. The main advan-
tage of the proposed approach is that it doesn’t require the user to ask a specific query
but let them concentrate on collecting observations of their interest. For a collected set of
observations, the system fulfills an abduction to suggest probable hypotheses and rate the
uncertainty of each hypothesis. We believe this way of using forensic visual surveillance
There are, however, still several open issues such as how to accurately define context
rules so that the derived hypotheses make sense, how to scale up the proposed approach
to large scale data, etc. Among them, as shown in the case study, we believe thorough
domain engineering and rule derivation is very critical because the final result would be
affected by this aspect. Nevertheless, we believe the proposed system has benefit and
unique property in the way of composing query especially in the forensic situation that the
possible.
Chapter 8
221
222
9 Conclusion
main contributions of this dissertation, highlight open issues and discuss extensions to this
work.
system architecture design as shown in Figure 9.1. Especially, we focused on the reuse of
logic programming and traditional databased techniques). Figure 9.1 - a) shows the key
Based on the proposed architecture, it has been extended to enable advanced reason-
ing aspects such as ‘default reasoning’ (in Chapter 5), ‘vague rule modeling’ (in Chapter 6),
Chapter 8). Because these topics were dealt mainly in logical reasoning sense, these chap-
ters correspond to the shaded area shown in Figure 9.1 - b). Especially, for the diagnostic
223
Contents
Figure 9.1: The Architectural Bic Picture and Topic Roadmap in terms of Chapters of this
Dissertation.
surveillance scenarios. Such rule sets can be regarded as domain knowledge. Therefore,
The main advantage of the presented approach is in the ‘flexibility’ of representing and
resolving epistemic contextual semantics by leveraging logic programming based data rep-
resentation model (this is due to the ‘modularity’ as explained in Section 2.3.3). Based on
the ‘flexible’ knowledge ‘expressive’ power, we have bestowed advanced reasoning features.
by our proposed approach is summarized in Table 9.1 (see Section 2.4.3 for details). The
table shows that the coverage of our subjective logic based approach is most broad. While
224
Contents
Figure 9.2: Positioning Subjective Logic among Other Uncertainty Representation For-
malisms.
some of previous work support advanced features such as ‘default reasoning’, ‘vague rule
modeling’ and ‘belief revision’, some features such as ‘bidirectional inference’ and ‘diagnos-
tic abduction’ are only supported by our approach. To cope with such ‘non-monotonic’
formalism. However, subjective logic itself is remain within the expressive power of propo-
sitional logic as of may ‘intensional’ approaches. In this dissertation to cope with complex
semantics, we adopted ‘predicate’ logic that is based on first-order logic. For the flexible
manipulation of facts and predicates on the need of arithmetic calculation, etc., we also
The discussion in Section 5.9.1 shows that our approach using subjective logic is ‘log-
ically’ robust. Especially, compared with bilattice based approach that can be considered
as Fuzzy-Belnap, Discussion 1 shows that our approach is free from the ‘paradoxes in the
truth table of multivalued logics’. Discussion 2 shows that the ‘logical soundness’ is impor-
Chapter 9
225
Contents
Knowledge Rule Based Ontology Rule Based Rule Based Rule Based Rule Based
Modeling (Prop. L) (DL) (FOL) (FOL) (Prop. L) (FOL)
Table 9.1: The Coverage of the Proposed Approach in terms of Reasoning Power.
tant in visual surveillance system. It makes the use of subjective logic attractive. In the
sense that the Fuzzy-Belnap has similar uncertainty representation tuple, it resembles sub-
jective logic. In addition to this, in Section 3.1.2, we have shown that the subjective logic
is derived from Dempster Shafer belief theory. Similarly, In Section 3.1.3, we have reviewed
that subjective logic is a special case of Bayesian, that uses Beta family of conjugated dis-
tribution among other distributions. Figure 9.2 shows our view of positioning subjective
logic among other uncertainty formalism. The reason for the complementary area between
subjective logic and Bayesian is because subjective logic can not handle other prior distri-
butions that can not be approximated to beta distribution. Naturally, subjective logic can
some of advanced techniques in Bayesian approach such as MCMC (Markov Chain Monte
Carlo) [33] is not captured in subjective logic theory. However, the fact that it bridges
multiple uncertainty formalism gives us a good intuition for interpreting and operating
226
Contents
For contradictory and imprecise knowledge based reasoning, we could benefit from
our subjective logic based ‘default reasoning’. Especially, in the forensic sense of visual
an incident or a report, it is natural to examine all positive and negative contextual evi-
dences that are related to the given hypothesis and fuse them to derive plausible conclusion
The support of ‘vague rules’ allows for representing lack of information about a
proposition. Therefore, we could roughly assign our subjective priors with lack of informa-
tion, and observations can also be represented with any degree of ignorance, therefore we
believe this better reflects human intuition and real world situations. Especially, rule can
embed its own opinion calculation scheme thereby, allows for sophisticated propagation of
how the reasoning results from uncertain spatio-temporal rules could be used with default
reasoning.
ditionals. The advantage is that it offers more choices to handle the trade-offs between
expressive power, semantic clarity and computational efficiency. For highly varying and
complex compositional events in visual surveillance data, we could benefit from the expres-
sive power and modularity of the extensional approach, while leveraging the intensional
Finally, the main advantage of the ‘diagnostic logical abduction’ is that it doesn’t
require the user to ask a specific query but let them concentrate on collecting observations
of their interest. For a collected set of observations, the system fulfills an abduction to
suggest probable hypotheses and rate the uncertainty of each hypothesis. We believe
this way of using forensic visual surveillance will offer the user more choices on semantic
analysis.
There are, however, drawbacks as well. One of the biggest drawbacks is the ‘scala-
bility’ against large scale data. To be fully applicable for practical real applications, the
Chapter 9
227
Contents
reasoning power should be ‘scalable’. Section 4.6 shows performance benchmark of triple
stores and rule-engines against different scale of metadata (i.e., fact base in view of rule-
engines). Unfortunately, it seems that logical formalism based approaches and the triple
stores of ontological metadata are by themselves not sufficiently scalable. The ‘scalability’
problem is common issue in the realm of ontology related research field and also expert
system related researches. One good news, however, is that there have been undergoing
active research focus on the ‘scalability’ issue. We have briefly introduced some possible
is also important and is an active research topic in the realm of high dimensional vector
indexing field. Therefore, we believe and hope that we could benefit from those researches
Besides the benefits and drawbacks discussed in the previous section, there are still open
While the proposed approach offers ‘flexibility’ and ‘expressive power’ for knowledge
modeling, the actual ‘knowledge acquisition’ and ‘expression’ should be done by ‘domain
experts’. Therefore, how to accurately define context rules so that the derived reason-
ing results make sense is an important issue. As shown especially in the case study of
Section 8.7, we believe thorough domain engineering and rule derivation is very critical
This is also related to an important issue on the verification of the proposed ap-
generate and share large scale video data set that contain rich ‘epistemic’ semantics. For
the verification and comparison, the video data set should also come with ground truth.
However, unlike data sets for traditional vision analytics such as human or object detection
and tracking, such ‘epistemic’ semantic data set is hard to generate in that : 1) it would
228
Contents
naturally contain not only multiple objects but also multiple human instances who should
regally agree on the use of the data set. 2) it would also naturally deal distributed multiple
camera that each instance should be labeled properly across the camera sources. 3) the
scenarios should be throughly considered before we build up such data set. Therefore, we
believe proper data set should be generated in a collaborative manner by many researchers
There are several open issues also related to subjective logic formalism such as how to
better model the reputational function, how to automatically assign proper prior opinions
to rules, how to extend this concept to multinomial frame based event, etc. Especially,
Therefore, our future research will cover above mentioned issues and extend the
shown approach to more complicated scenarios using automatically generated large scale
data.
We believe that there are several important contributions in this dissertation. First, this
dissertation lays out a systematic approach and methodical support for a more intelligent
semantic reasoning system, especially in the sense of ‘epistemic’ forensic (-post) analysis
based retrieval is another remarkable approach that tries to search a scene that matches
to given image queries. Such methods have been limited to handle only simple and de-
terministic queries due to their heavy dependency on signal level processing. Only a few
researches have been attempted to handle complex semantic aspects such as using tradi-
tional database and ontology technics. However, such work lacks handling uncertainty and
heavily rely on deterministic data scheme thereby, the data processing components are too
much tied on the pre-defined scheme. Even a small change hits the consistency of the sys-
Chapter 9
229
Contents
tem thereby, the change forces entire system to be modified. Unlike legacy approaches, this
work proposes the use of logic programming that is extended with subjective logic theory
for uncertainty handling and epistemic reasoning. To the best of our knowledge, this is the
first attempt that subjective logic theory is applied to visual surveillance application.
It is worth to note that one remarkable research has been attempted in a simi-
lar manner using bilattice framework implemented with prolog. The bilattice framework
demonstrated reasoning power such as belief revision and default reasoning. However, it is
important to note that the bilattice framework based approach lacks soundness in terms of
logic theory. Indeed, the operations in bilattice framework are rather related to fuzzy the-
ory than traditional probability based Bayesian. Most of the vision analytic modules today
tend to based on Bayesian approach throughout its design and training phase. Therefore,
for system engineering to reuse such analytic data, providing a framework that can be also
reasoned in a Bayesian sense is preferred. In contrast, subjective logic not only provides
Bayesian sense of interpreting reasoning data but also bridges many aspects of bilattice
based approach, fuzzy based approach, dempster shaferian approach and Bayesian ap-
proach. Especially, we have demonstrated how conditional premises can be factorized and
handled in a Bayesian sense by using subjective logic’s deduction and abduction operators.
tion by introducing a reputation concept. We believe this way, we better capture human
intuition on modeling and handling vague propositions. We further extend our approach
to model default reasoning behaviour. To the best of our knowledge, this is also the first
time on modeling default reasoning in a schematic way using subjective logic operators.
A comprehensive comparison with L-fuzzy logics and billatice theory is conducted. This
comparison shows many overlap between subjective logic theory and bilattice in the behav-
showed that subjective logic is more robust in handling logical premises. This is again due
to the lack of soundness in bilattice that often derives counter intuitive reasoning result
230
Contents
Based on the proposed approach, we further extend the system so that it can handle
abductive reasoning. Semantically speaking, such abductive queries are more complex to
handle due to its ambiguity. We tackled this by adding a layer that reason about queries to
be asked and a layer that uses abductive logic programming to draw potential hypotheses
given set of observations. Our demonstration showed the feasibility of using the proposed
approach on diagnostic queries. To the best of our knowledge, this is also the first time
The attempt to semantic reasoning done in this dissertation has given several insights
into the design of intelligent visual surveillance system. Development of better intelligent
visual surveillance systems will require feedback. That is, deriving machine-processable
reasoning categories to develop a semantic reasoning mechanism and the systems are often
made by using the insights during the attempt to building such a system. In fact, this
ferent uncertainty formalisms. Preliminary results indicate that this approach is feasible
Chapter 9
231
232
List of Figures
3.4 Linguistic Fuzzy Category for Opinion Triangle as a Function of the Base
Rate [139]. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
3.6 Examples of Applying Bayes’ Theorem to Different Priors and Given a Like-
lihood. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
tem. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
233
List of Figures
Ontology. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
4.7 Illustrative Example of VA-file : a) VA-file for 2-D Vectors and b) Lower
and Upper Bound between a Given Query Vector q and Feature Vectors in
a Cell. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
4.10 Flat-file Style Database Tables in ERD (Entity Relation Diagram) Each
4.11 Querying and Retrieval Graphical User Interface and Retrieval Results of
4.12 Benchmark Result in the Work of Jürgen Bock et al. [30]. The Figures
show the Average Load and Query Time with a Particular Reasoner on the
Dataset shown in Table 4.3. The ‘◦’ indicates Time-out (> 5 min). . . . . . 89
234
List of Figures
Database Models. The Figures show the Average Response Time on Dif-
ferent Datasets shown in Table 4.5. The ‘◦’ indicates Time-out (> 1200
sec). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
5.1 Examples of Lattice and Bilattices (a) a Lattice (b) Bilattice correspond-
also can be seen as LI Triangle for Interval-Valued Fuzzy Logic), and (b)
5.4 (a) Belnap Logic, FOUR (b) Default Logic (c) Multivalued (Prioritized)
5.5 (a)The Bilattice and Opinion Triangle Space for Belnap FOUR, (b) The
Bilattice and Opinion Triangle Space for Default Logic, (c) The Bilattice
5.6 Reasoning Results of Example 3 in Opinion Space and L2 & I ∗ (L). . . . . . 133
5.7 Reasoning Results of Example 4 in Opinion Space and L2 & I ∗ (L). . . . . . 137
5.9 Reasoning Results of Example 5 in Opinion Space and L2 & I ∗ (L). . . . . . 143
5.11 (a) Two Examples of U ∧ ⊥ (b) Two Examples of U ∨ ⊥ in Subjective Logic. 148
5.12 Reasoning Results of Modified Example 3 in Opinion Space and L2 & I ∗ (L). 153
6.1 Vague Rule Modeling using Subjective Logic’s Reputation Operator. . . . . 164
235
List of Figures
6.4 Samples of Reputational Opinion according to Distance and Equation 6.1. . 168
7.4 Illustration of Abductive Conditional Reasoning for Case Study I (see http:
//persons.unik.no/josang/sl/Op.html). . . . . . . . . . . . . . . . . . . 194
7.5 Screen Captures demonstrating Belief Revision on the Proposition ‘the es-
7.6 Scenario Setting for Case Study 2. - Two Cars, C1 and C2. An Empty
7.8 Illustration of Abductive Conditional Reasoning for Case Study 2 (see http:
//persons.unik.no/josang/sl/Op.html). . . . . . . . . . . . . . . . . . . 205
8.2 The Setup for Case Study Scenario with 3 Cameras Deployed. . . . . . . . . 216
9.1 The Architectural Bic Picture and Topic Roadmap in terms of Chapters of
malisms. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 225
236
List of Tables
2.1 Formal Logics and Their Ontological and Epistemological Commitments [146]. 21
4.1 Metadata Vocabulary and the Property of Each Primitive Semantic Item. . 82
4.3 Statistics of Test Ontologies used in the Work of Jürgen Bock et al. [30].
Relation R(a,b)). The Red Boxed Row Shows the Statistics of Our Manual
poser. The Red Arrow Indicates that Our Annotation Resembles wine_1
5.1 Truth Table of Bilattice Operators on (a) Belnap (b) Default Logic (The
237
List of Tables
5.2 An Example of Applying Dempster’s Rule and Consensus Operator for Un-
5.3 An Example of Applying Dempster’s Rule and Consensus Operator for Dog-
matic Belief Fusion - Dempster’s Rule Deriving Counter Intuitive Result [94].116
9.1 The Coverage of the Proposed Approach in terms of Reasoning Power. . . . 226
238
Bibliography
[1] CLIPS - Basic Programming Guide - NASA JSC-25012. NASA. 57, 68, 78, 91, 186,
214
[5] Rushes - european research project on multimedia search and retrieval of rushes data.
http://www.rushes-project.eu/. 1
[9] A normal approximation for beta and gamma tail probabilities. Probability Theory and
[10] Report of the Official Account of the Bombings in London on 7th July 2005. House of
[12] J. K. Aggarwal and Q. Cai. Human motion analysis: A review. pages 90–102, June
1997. 31
239
Bibliography
[13] U. Akdemir, P. Turaga, and R. Chellappa. An ontology based approach for activity
Multimedia, ACM MM ’2008, pages 709–712. ACM, 2008. 34, 35, 103, 211, 226
[14] H. T. Al-Feel, M. Koutb, and H. Suoror. Semantic web on scope : A new architectural
model for the semantic web. Journal of Computer Science, 4(7):613–624, 2008. 22, 23,
233
[16] A. Altadmri and A. Ahmed. Automatic semantic video annotation in wide domain
Modeling human activity from voxel person using fuzzy logic. IEEE Transactions on
Fuzzy Systems, 17(1):39–49, February 2009. 34, 35, 36, 103, 162, 211, 226
Approaches to Reasoning with Uncertainty, ECSQARU’ 2005, LNAI 3571, pages 563–
[21] K. T. Atanassov. Intuitionistic fuzzy sets. Fuzzy Sets and Systems, 20(1):87–96,
240
Bibliography
[23] K. T. Atanassov. Intuitionistic fuzzy sets: past, present and future. In Proceedings
of the 3rd Conference of the European Society for Fuzzy Logic and Technology, pages
Description Logic Handbook. Cambridge Univ. press., 2005. 20, 22, 34, 60
and recognition for semantic annotation of video. Multimedia Tools and Applications,
51:279–302, 2011. 1
[26] S. Benferhat, A. Saffiotti, and P. Smets. Belief functions and default reasoning. Arti-
and videos for multimedia analysis. In Proceedings of the 2005, 2nd European Semantic
[29] A. Bobick and J. Davis. The recognition of human movement using temporal tem-
2001. 31
[30] J. Bock, P. Haase, Q. Ji, and R. Volz. Benchmarking owl reasoners. In Proceedings
ARea ’2008, June 2008. 87, 88, 89, 90, 234, 237
241
Bibliography
[32] B. Bolles and R. Nevatia. A hierarchical video event ontology in owl. In ARDA
[36] B. G. Buchanan and E. H. Shortliffe. Rule Based Expert Systems: The MYCIN
[37] H. Buxton and S. Gong. Visual surveillance in a dynamic and uncertain world. Journal
[39] C. L. Chang and A. Walker. Prosql: a prolog programming interface with sql/ds.
In Proceedings from the first international workshop on Expert database systems, pages
233–246, Redwood City, CA, USA, 1986. Benjamin-Cummings Publishing Co., Inc. 95
[40] E. F. Codd. Redundancy, and consistency of relations stored in large data banks.
[41] E. F. Codd. A relational model of data for large shared data banks. Communications
[42] C. Cornelis, K. T. Atanassov, and E. E. Kerre. Intuitionistic fuzzy sets and interval-
valued fuzzy sets: a critical comparison. In Proceedings of the 3rd Conference of the
242
Bibliography
European Society for Fuzzy Logic and Technology, pages 159–163, September 2003. 102,
104, 107
and Beyond Essays in Honour of Robert A. Kowalski (LNAI Volumes 2408), pages
groups of people, and crowd. IEEE Seminar Intelligent Distributed Surveillance Sys-
[47] R. Datta, D. Joshi, J. Li, James, and Z. Wang. Image retrieval: Ideas, influences, and
[48] J. Davies. Semantic Web Technologies: Trends and Research in Ontology-based Sys-
[49] H. M. Dee and S. A. Velastin. How close are we to solving the problem of automated
logic: Logic programming and beyond. Lecture Notes in Computer Science, 2407:402–
[51] M. Denecker and D. D. Schreye. Sldnfa: an abductive procedure for normal abductive
243
Bibliography
[52] N. Dimitrova, H.-J. Zhang, B. Shahraray, I. Sezan, T. Huang, and A. Zakhor. Appli-
cations of video-content analysis and retrieval. IEEE MultiMedia, 9:42–55, July 2002.
[54] A. Dorado, J. Calic, and E. Izquierdo. A rule-based video annotation system. IEEE
Transactions on Circuits and Systems for Video Technology, pages 622–633, 2004. 35,
226
[55] C. Draxler. A powerful prolog to sql compiler. Technical report, Center for Information
95
214
on Computer Vision, ICCV ’1998, pages 350–355, Washington, DC, USA, 1998. IEEE
Computer Society. 33
[59] G. Ferrin, L. Snidaro, I. Visentini, and G. L. Foresti. Abduction for human behav-
ior analysis in video surveillance scenarios. In Proceedings of the 2008 1st Interna-
tional Workshop on Tracking Humans for the Evaluation of their Motion in Image Se-
244
Bibliography
[61] M. Fitting. Kleenes three-valued logics and their children. Fundamental Informaticae,
[62] P. Flach. On the state of the art in machine learning : A personal review. Artificial
[63] M. Fleischman, P. Decamp, and D. Roy. Mining temporal patterns of movement for
[64] C. Forgy. Rete: A fast algorithm for the many pattern/many object pattern match
1995. 1
[66] J. Gallier. Chapter 9 sld resolution and logic programming. In Logic for Computer
[68] M. Gehrke, C. Walker, and E. Walker. Some comments on interval-valued fuzzy sets.
intelligence. Computational Intelligence, 4(3):256–316, 1988. 34, 100, 104, 106, 111,
211, 235
245
Bibliography
[71] S. Gong and J. Ng. Learning pixel-wise signal energy for understanding semantics. In
Proceedings of the 2001, British Machine Vision Conference, pages 1183–1189, 2001.
31
[72] S. Gong and T. Xiang. Recognition of group activities using dy- namic probabilistic
[73] S. Gottwald and P. Hajek. Triangular norm based mathematical fuzzy logic. In E. P.
Klement and R. Mesiar, editors, Logical, Algebraic, Analytic and Probabilistic Aspects
[74] V. Gouaillier. Intelligent video surveillance: Promises and challenges. Technical re-
port, 2009. 2, 5
[75] T. Gruber. Towards principles for the design of ontologies used for knowledge sharing.
[76] M. Guinness and V. Harmelen. OWL Web Ontology Language Overview. W3C Rec-
[77] R. Hähnle. Many-valued logic, partiality, and abstraction in formal specification lan-
[78] S. Han, A. Hutter, and W. Stechele. Toward contextual forensic retrieval for visual
Proceedings of the 2011 IEEE International Conference on Multimedia and Expo, ICME
[80] S. Han, B. Koo, A. Hutter, V. Shet, and W. Stechele. Subjective logic based hybrid
246
Bibliography
of the 2010 7th IEEE International Conference on Advanced Video and Signal Based
Surveillance, AVSS ’2010, pages 337–344, 2010. 35, 36, 38, 157, 211, 226
[81] S. Han, B. Koo, A. Hutter, and W. Stechele. Forensic reasoning upon pre-obtained
Proceedings of the 2010 11th International Workshop on Image Analysis for Multimedia
Interactive Services, WIAMIS’ 2010, 2010. 35, 36, 38, 157, 211, 226
[82] S. Han, B. Koo, and W. Stechele. Subjective logic based approach to modeling default
reasoning for visual surveillance. In Proceedings of the 2010 IEEE 4th International
Conference on Semantic Computing, ICSC ’2010, pages 112–119. IEEE Computer So-
ciety, 2010. 35, 36, 38, 69, 173, 211, 219, 226
[83] S. Han and W. Stechele. Default reasoning for forensic visual surveillance based on sub-
jective logic and its comparison with l-fuzzy set based approaches. International Journal
[84] P. Hayes. On semantic nets, frames and associations. In Proceedings of the 1977
[85] S. Hongeng, R. Nevatia, and F. Bremond. Running head: Video-based event recogni-
tion methods. Journal of Computer Vision and Image Understanding - Special number
[86] W. Hu, T. Tan, L. Wang, and S. Maybank. A survey on visual surveil- lance of
object motion and behaviors. IEEE Transactions on System, Man, and Cybernetics,
34(3):334–352, 2004. 31
247
Bibliography
296–306, 1984. 95
[90] F. V. Jensen. Bayesian Networks and Decision Graphs. Information Science and
Statistics, 2007. 32
[91] K. H. Jo, Y. Kuno, and Y. Shirai. Manipulative hand gesture recognition using task
knowledge for human computer interaction. In Proceedings of the 1998, 3rd Interna-
[92] A. Jøsang. Artificial reasoning with subjective logic. In Proceedings of the 1997 2nd
Australian Workshop on Commonsense Reasoning, 1997. 37, 48, 100, 101, 211
Fuzziness and Knowledge-Based Systems, 9(3):271–311, 2001. 10, 36, 37, 38, 39, 40, 41,
[94] A. Jøsang. The consensus operator for combining beliefs. Artificial Intelligence Jour-
nal, 38(1):157–170, 2006. 38, 39, 40, 114, 116, 118, 149, 150, 155, 169, 192, 202, 233,
237, 238
Logic and Soft Computing, 15(1):5–38, 2008. 51, 52, 182, 183
[96] A. Jøsang. Cumulative and averaging unfusion of beliefs. In Proceedings of the Inter-
Journal of Approximate Reasoning, 142(1-2):19–51, 2004. 52, 123, 149, 150, 154, 169,
185, 191
248
Bibliography
tion Fusion, pages 1133–1140, 2003. 38, 48, 50, 114, 149, 150
[99] A. Jøsang, S. Marsh, and S. Pope. Exploring different types of trust propagation.
to Reasoning with Uncertainty, ECSQARU’ 2005, Barcelona, Spain, July 2005. 51, 52,
[101] A. C. Kakas, A. Michael, and C. Mourlas. Aclp: Abductive constraint logic pro-
[102] G. Lavee, M. Rudzsky, E. Rivlin, and A. Borzin. Understanding video events: A sur-
report, Technion - Israel Institute of Technology, Haifa, Israel, Tech. Rep. CIS-2009-06,
2009. 31
[103] R. Lee. Scalability report on triple store applications. Technical report, MIT, July
2004. 61
[104] J. Lloyd and W. John. Foundations of logic programming. Springer-Verlag New York,
Inc., 1987. 61
[107] R. Ludwig. Kant für Anfänger - Die Kritik der reinen Vernunft : Eine Lese -
[108] F. Lv and R. Nevatia. Single view human action recognition using key pose matching
249
Bibliography
and viterbi path searching. In Proceedings of the 2007, IEEE International Conference
on Computer Vision and Pattern Recognition, CVPR ’2007, pages 1–8, 2007. 32
[109] F. Lv, X. Song, B. Wu, V. K. Singh, , and R. Nevatia. Left-luggage detection using
Performance Evaluation in Tracking and Surveillance, PETS ’2006, pages 83–90, 2006.
32
[110] J. Ma, W. Liu, P. Miller, and W. Yan. Event composition with imperfect information
for bus surveillance. In Proceedings of the 2009 6th IEEE International Conference
on Advanced Video and Signal Based Surveillance, AVSS ’2009, pages 382–387. IEEE
[111] L. Magnani. Abduction, Reason, and Science. Processes of Discovery and Explana-
and H. Uchiyama. Efficient integration of prolog and relational databases in the ned
[113] D. Makris and T. Ellis. Spatial and probabilistic modelling of pedestrian behaviour.
[114] D. Makris and T. Ellis. Learning semantic scene models from observing activity
35(3):397–408, 2005. 33
250
Bibliography
and analysis from video streams. IEEE Transactions on Pattern Analysis and Machine
Psychology of Computer Vision, pages 211–277. McGraw Hill, San Mateo, CA, 1975.
20, 21
the recognition of human behaviors from video. In Proceedings of the 2003, Interna-
tional Conference on Computer Vision Systems, ICVS’ 2003, pages 68–77, 2003. 33
[121] J. Muncaster and Y. Ma. Activity recognition using dynamic bayesian networks with
automatic state selection. In Proceedings of the 2007, IEEE Workshop on Motion and
[122] R. Nevatia, J. Hobbs, and B. Bolles. An ontology for video event representation.
In Proceedings of the 2004 IEEE Workshop on Event Detection and Recognition, June
2004. 24
[123] NIST. Trec video retrieval evaluation - digital video retrieval at nist. http://
trecvid.nist.gov/. 229
[124] S. Nowozin, G. Bakir, and K. Tsuda. Discriminative subsequence mining for action
Proceedings of the 2001 6th International Conference on Logic Programming and Non-
monotonic Reasoning, LPNMR’ 2001, pages 393–396. Springer Verlag, 2001. 213, 214
251
Bibliography
tion of pose sequences for action recognition. In Proceedings of the 2004 Workshop on
[128] N. Oliver, E. Horvitz, and A. Garg. Layered representations for human activity recog-
hash function types and querying mechanisms. Pattern Recogn. Letter, 31(11):1348–
ence. Morgan kaufmann, San Mateo, CA, 1988. 2, 4, 9, 25, 27, 100, 103, 180, 181,
233
[131] C. S. Peirce. Philosophical Writings of Peirce. Dover Publications, New York, 1955.
2007. 212
dia low-level features. Lecture Note in Computer Science, Knowledge based intelligent
[134] K. Phillips-Fein. The 9/11 Commission Report - Final Report of the National Com-
mission on Terrorist Attacks Upon the United States. Official Government Edition,U.S.
252
Bibliography
[135] C. Piciarelli, G. Foresti, and L. Snidaro. Trajectory clustering and its applications
[136] M. Pittore, C. Basso, and A. Verri. Representing and recognizing visual dynamic
[138] J. L. Pollock and A. S. Gillies. Belief revision and epistemology. Synthese - Interna-
2000. 3, 9
[139] S. Pope and A. Jøsang. Analsysis of competing hypotheses using subjective logic.
[140] S. Pope and A. Jøsang. Abductive reasoning with uncertainty. In Proceedings of the
[143] R. Reiter. A logic for default reasoning. Artificial Intelligence, 13:81–132, 1980. 9,
[145] P. Ribeiro, P. Moreno, and J. S. Victor. Boosting with temporal consistent learners:
253
Bibliography
[146] S. Russell and P. Norvig. Artificial Intelligence: A Modern Approach. Prantice Hall,
retrieval engine for multimedia semantic units. Multimedia Tools and Applications,
[148] N. Shachtman. London bomb plot: Surveillance web slips again? http://www.
of the 2005 IEEE Conference on Computer Vision and Pattern Recognition, CVPR’
[152] V. D. Shet, D. Harwood, and L. S. Davis. Vidmap: video monitoring of activity with
Video and Signal Based Surveillance, AVSS’ 2005, pages 224–229, 2005. 34, 35, 103,
[153] V. D. Shet, D. Harwood, and L. S. Davis. Multivalued default logic for identity
Computer Vision, ECCV’ 2006, pages 119–132, 2006. 102, 103, 104, 130, 131, 134, 211
254
Bibliography
on Video Surveillance & Sensor Networks, pages 79–86, 2006. 102, 103, 104, 130, 134
ing for human detection. In Proceedings of the 2007 IEEE International Conference on
Computer Vision and Pattern Recognition, CVPR’ 2007, pages 1–8, 2007. 34, 35, 104,
226
ontology for multimedia reasoning. In Proceedings of the 2005 6th International Work-
shop on Image Analysis for Multimedia Interactive Services, WIAMIS’ 2005, Montreux,
[157] P. Smith, N. da Vitoria Lobo, and M. Shah. Temporal boost for event recognition.
[159] R. Stalnaker. Probability and conditionals. The University of Western Ontario Series
182
[160] G. Stoilos, G. B. Stamou, and J. Pan. Fuzzy extensions of owl: Logical properties and
51(6), 2010. 23
255
Bibliography
[163] Y. Sure, S. Staab, and R.Studer. Methodology for development and employement of
of human activities: A survey. IEEE Transactions on Circuits and Systems for Video
generative (tdrgbn) network for gesture based control. In Proceeding of the International
[168] W3C. SPARQL Query Language for RDF: W3C Working Draft. http://www.w3.
[169] W3C. Notation3 - a readable language for data on the web. http://www.w3.org/
Designnumbers/Notation3.html, 2000. 73
[171] R. Weber, H. Schek, and J. Blott. A quantitative analysis and performance study
[172] T. Weithöner, T. Liebig, M. Luther, and S. Böhm. Whats wrong with owl bench-
256
Bibliography
reasoning with owl. In Proceedings of the 4th European Conference on The Semantic
Web: Research and Applications, ESWC ’2007, pages 296–310, Berlin, Heidelberg, 2007.
Springer-Verlag. 87
Reasoner. 74, 87
[176] L. A. Zadeh. Fuzzy sets. Information and Control, 8:338–353, 1965. 20, 100, 107
[177] L. A. Zadeh. Outline of a new approach to the analysis of complex systems and
[179] L. A. Zadeh. Is there a need for fuzzy logic? Information Sciences, 178(13):2751–
[180] L. Zelink-Manor and M. Irani. Statistical analysis of dynamic actions. IEEE Trans-
[181] F. Zhang, Z. M. Ma, J. Cheng, and X. Meng. Fuzzy semantic web ontology learning
from fuzzy uml model. In Proceeding of the 18th ACM conference on Information and
knowledge management, CIKM ’2009, pages 1007–1016, New York, NY, USA, 2009.
ACM. 23
257
258