0% found this document useful (0 votes)
37 views6 pages

Emotion Prediction As Computation Over A Generative Theory of Mind

Uploaded by

im.lifengfan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
37 views6 pages

Emotion Prediction As Computation Over A Generative Theory of Mind

Uploaded by

im.lifengfan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

Iterative Theory of Mind Assay of Multimodal AI Models

Rohini Elora Das 1 Rajarshi Das 2

Abstract without controversy. Within the AI community, there is on-


going debate about whether these highly proficient language
The concept of artificial general intelligence
models genuinely understand language and the physical and
(AGI) has sparked intense debates across vari-
social contexts it encodes in a humanlike manner (Mitchell
ous sectors, fueled by the capabilities of Large
& Krakauer, 2023; Messeri & Crockett, 2024). Addition-
Language Model-based AI systems like ChatGPT.
ally, questions persist about their ability to reason, plan, and
However, the AI community remains divided on
make decisions based on limited experiences (Mitchell et al.,
whether such models truly understand language
2023).
and its contexts. Developing multimodal AI sys-
tems, which can engage with the user in multiple
input and output modalities, is seen as a crucial
step towards AGI. We employ a novel iterative
Theory of Mind (iToM) test approach to reveal
limitations of current multimodal LLMs like Chat-
GPT 4o in converging to coherent and unified
internal world models which results in illogical
and inconsistent user interactions both within and
across the different input and output modalities.
We also identify new multimodal confabulations
(“hallucinations”), particularly in languages with
less training data, such as Bengali.
Figure 1. (a) To the user, a multimodal AI system may appear
as a single AI system with its own unified Theory of Mind.
1. Introduction (b) Schematic diagram of a typical multimodal AI system with
The concept of “artificial general intelligence” (AGI) has different internal LLM modules, potentially with their own Theory
of Mind, that are fused together to serve different input and output
ignited widespread fascination across business, government,
modalities.
and media sectors, prompting vigorous debates about its im-
plications and potential consequences (Mitchell, 2024). Re-
cent advancements in Large Language Model (LLM)-based Inspired by the human brain’s ability to perceive and in-
AI systems, such as ChatGPT, have significantly fueled this terpret information through various senses (Constantinescu
excitement. These systems have provided unprecedented op- et al., 2016), the development of multimodal AI systems is
portunities for individuals worldwide to engage with AI on seen as a critical step toward achieving AGI. Multimodal
a deeply personal level through natural language interfaces. AI systems, such as OpenAI’s GPT-4o (omni), can pro-
Remarkable claims about LLMs approaching or surpassing cess and integrate multiple types of data inputs, including
human-level performance across various tasks have inten- text, images, audio, and video. These systems leverage a
sified the pursuit of AGI among leading tech companies. combination of AI techniques, including natural language
Achieving and demonstrating AGI is now seen as a logical processing, computer vision, and speech recognition, to
progression in the AI field. However, this pursuit is not understand and generate content across different formats.
By employing data fusion techniques, they align informa-
1
New York University, New York, NY, USA 2 MQube Cog- tion from diverse sources, resulting in an architecture that
nition, New York, NY, USA. Correspondence to: Rohini Das combines input, fusion, and output modules to deliver pre-
<[email protected]>. dictions or generate content. To users, these multimodal
Proceedings of the 41 st International Conference on Machine AI systems appear as cohesive entities capable of under-
Learning, Vienna, Austria. PMLR 235, 2024. Copyright 2024 by standing, reasoning, and communicating holistically across
the author(s). different modalities (Figure 1a).

1
Iterative Theory of Mind Assay of Multimodal AI Models

Figure 2. Iterative Theory of Mind (ToM) assay of multimodal AI systems.

To systematically investigate whether these systems truly in unison, these tests would produce results that are coherent
understand the content they generate, we adopt an approach and consistent both within each iteration and across multiple
from experimental psychology involving Theory of Mind iterations. Human performance on the same tasks across
(ToM) (Wang et al., 2024). Theory of Mind—the ability multiple iterations serves as a benchmark for comparison.
to track one’s own mental state or other people’s mental
The MANAS framework allows for tests to be conducted
states—is a fundamental aspect of human cognition. It has
in the language chosen by the user where the prompts and
been used to compare human and LLM performance on
responses in text and audio was constrained to be in the
comprehensive measures of understanding (Strachan et al.,
selected language. All tests were performed in both English
2024). We extend this concept to explore ToM within and
and Bengali. The same sequence of tests were also offered
across modalities in multimodal AI systems, providing a
to human subjects (high-school freshmen and former SFI
deeper understanding of their capabilities and limitations
Complex System Summer School students).
(Figure 1b).
Preliminary results from our MANAS (Theory of Mind 3. Results
Assay of Natural and Artificial Intelligent Systems) show
that current multimodal AI systems like GPT-4o are lim- Sample results from MANAS are presented below.
ited in their ability to create an integrated, unified, and
coherent world model from the different internal modules
serving various modalities. Additionally, MANAS helped
to uncover a new type of multimodal confabulation (or ”hal-
lucination”) in languages with relatively limited training
data compared to English. For example, in Bengali, the
seventh most spoken language in the world with 272 million
speakers, GPT-4o can communicate through text. But it also
generates confabulating images of scripts that might appear
as Bengali alphabets to non-native Bengali readers. Such
aberrant behavior has not been observed at such scale in
English.

2. Theory of Mind Assay of MultiModal Chat


GPT4o
The key idea behind MANAS is to evaluate Multimodal
Figure 3. Task 1. Iteration 1. Image output from a text prompt.
AI systems by employing Theory of Mind tests iteratively, Note the six-sided dice has seemingly three sides of five black dots
switching prompts and responses between different input (pips). Also, note the mismatch between the dice and its reflection
and output modalities (Figure 2). In a Multimodal AI system in the mirror (five pips on the top face of the dice and four pips in
where internal LLM modules are well-integrated and work the reflection).

2
Iterative Theory of Mind Assay of Multimodal AI Models

Figure 4. Task 1. Iteration 2. Text output from the image prompt


Figure 6. Task 1: Iteration 4. Text output from the image prompt
in Figure 3. Note the discrepancy between the text description and
in Figure 5. Note the discrepancy between this text description and
the image output in Figure 3.
the image output in Figure 5, text output in Figure 4, and image
output in Figure 3.

Figure 5. Task 1. Iteration 3. Image output from the text prompt


in Figure 4. Note the discrepancy between this figure and the text
description in Figure 4 as well as the image output in Figure 3.

1. Prompt (text) ChatGPT4o to imagine and sketch a reg-


ular six sided dice in front of a mirror (Figure 3). Figure 7. Task 2: Iteration 1. Text description from an image of a
Lorenz attractor.
2. Save the generated image in Step 1 and use it as a
prompt (image) to ChatGPT4o, requesting a textual
description of the prompt (Figure 4).

3. Save the generated text in Step 2 and use it as a prompt


3.1. Task 2: Imagine the Lorenz attractor: Now draw
(text) to ChatGPT4o, requesting it to generate an image
and describe it
given description of the prompt (Figure 5).
Except for the initial image prompt of a Lorenz attractor,
4. Save the generated image in Step 3 and use it as a the sequence of steps in this task was identical to those
prompt (image) to ChatGPT4o, requesting a textual in Task 1. While ChatGPT 4o accurately recognized the
description of the prompt (Figure 6). image of the Lorenz attractor and consistently described its
three-dimensional structure (Figure 7), it failed to draw an
Comparing image and text outputs from Steps 1 through 4 approximate sketch based on those descriptions (Figure 8).
show continued inconsistencies both within each represen- Although ChatGPT 4o can write a Python program that
tation as well between responses generated from successive correctly draws the Lorenz attractor, it could not produce
steps. Results from human subects (high school freshmen the drawing correctly during our tests. Human subjects who
from the art class) (not presented here) receiving similar set remembered the shape of the Lorenz attractor were more
of instructions remained consistent over multiple iterations. consistent in their representations.

3
Iterative Theory of Mind Assay of Multimodal AI Models

Figure 8. Task 2: Iteration 2. Image output from text description Figure 10. Task 3: Iteration 1. Image of a directed graph used as a
from Figure 7. prompt for Figure 11.

Figure 11. Task 3: Iteration 1. Novel graph path simulation of a


graph used as a prompt in Figure 10.

Figure 9. Task 2: Iteration 2 (Bengali). Image output from Bengali


text description from Figure 7.

When faced with the same task in Bengali, ChatGPT 4o con-


sistently generated confabulated titles and labels in scripts
resembling Devanagari or Bengali (Figure 9). Upon closer
inspection, each of the Bengali characters was found to be
fake. Human subjects who read or write Bengali did not
make this mistake.

3.2. Task 3: Find and extract directed graph in an


image, generate sequence from the graph
Figure 12. Task 3: Iteration 2. Image output from text description
The task forces ChatGPT4o to contruct an internal model from Figure 11. Note the discrepancy between the image output
of a directed graph from an image prompt. This imagined from Figure 10 and this image output.
graph is then employed to generate sequences by traversing
3.3. Task 4: False Belief Test
paths in the graph. Future versions of this task will include
additional steps to test Multimodal AI systems to infer finite To explore ChatGPT-4o’s ability to track its own internal
state machines from generated sequences. model in representing other people’s mental states in dual or

4
Iterative Theory of Mind Assay of Multimodal AI Models

multi-agent environments, we adapt the False Belief ToM


test used in Clinical psychology for this task. In a False
Belief test, a character is exposed to partial information
leading to a ‘false belief’ in contrast to another character
who is exposed to the full information. Can ChatGPT-4o
represent and articulate these different beliefs consistently
across different modalities?
Except for the initial audio prompt to set the False Belief
test scenario, the sequence of steps in this task was identical
to those in Task 1 as presented in Figure 13, Figure 14, and
Figure 15.

Figure 14. Task 4: Iteration 2. A detailed description is generated


based solely on the output image from Figure 13 and follow-up
questions about the image are answered. In this example, the bag
is accurately described as containing both bananas and apples.

Figure 13. Task 4: Iteration 1. ChatGPT-4o is asked to depict an


image of a False Belief ToM scenario through an audio prompt. In
the generated image, bananas are added to a bag without heeding
the audio prompt. When asked about the contents of the bag, only
apples are mentioned, ignoring the bananas visible inside the bag.
While ChatGPT4o correctly identifies the person with the false
belief, the reason for the false belief is not identified accurately.

4. Conclusion and Future Work


In this work, we employed iterative Theory of Mind tests to
begin exploring ChatGPT-4o’s multimodal capabilities in
understanding and representing its own internal spatial and
abstract states, and in generating responses based on that
understanding across different modalities. When compared
to human performance benchmarks for the same tasks, it
is evident that GPT-4o is limited in its ability to create and Figure 15. Task 4: Iteration 3. The AI system generates an image
serve an integrated, unified, coherent and consistent model based on the description from Figure 13b and answers questions
of its own state from the different internal modules serving about the generated image. In this example, the system incorrectly
various modalities. In future work, we plan to conduct analyzes its own generated image, describing the girl in the red
more complex iterative Theory of Mind tests to investigate dress as “surprised and confused” based on her expression, even
though the image actually shows her with a pleased expression.
ChatGPT-4o’s ability to track other people’s mental states
multi-agent environments.

5
Iterative Theory of Mind Assay of Multimodal AI Models

References
Constantinescu, A. O., O’Reilly, J. X., and Behrens, T. E. J.
Organizing conceptual knowledge in humans with a grid-
like code. Science, 352(6292):1464–1468, 2016.

Messeri, L. and Crockett, M. J. Artificial intelligence and


illusions of understanding in scientific research. Nature,
627, 2024.
Mitchell, M. Debates on the nature of artificial general
intelligence. Science, 383(6689), 2024.
Mitchell, M. and Krakauer, D. C. The debate over under-
standing in ai’s large language models. Proceedings of
the National Academy of Sciences, 120(13), 2023.
Mitchell, M., Palmarini, A. B., and Moskvichev, A. Com-
paring humans, gpt-4, and gpt-4v on abstraction and rea-
soning tasks, 2023.
Strachan, J. W. A., Albergo, D., Borghini, G., Pansardi, O.,
Scaliti, E., Gupta, S., Saxon, K., Rufo, A., Panzeri, S.,
Manzi, G., Graziani, M., and Becchio, C. Testing theory
of mind in large language models and humans. Nature
Human Behaviour, 2024.
Wang, Q., Walsh, S., Si, M., Kephart, J., Weisz, J., and Noel,
A. Theory of mind in human-ai interaction. Extended
Abstracts of the CHI Conference on Human Factors in
Computing Systems, 2024.

You might also like