Robot Learning Theory of Mind through Self-Observation: Exploiting
the Intentions-Beliefs Synergy
Francesca Bianco1and Dimitri Ognibene2,1,∗
Abstract— In complex environments, where the human sen-
sory system reaches its limits, our behaviour is strongly driven
by our beliefs about the state of the world around us. Accessing
others’ beliefs, intentions, or mental states in general, could
thus allow for more effective social interactions in natural
contexts. Yet these variables are not directly observable. Theory
of Mind (TOM), the ability to attribute to other agents’ beliefs,
intentions, or mental states in general, is a crucial feature of
human social interaction and has become of interest to the
robotics community. Recently, new models that are able to
learn TOM have been introduced. In this paper, we show the
synergy between learning to predict low-level mental states,
such as intentions and goals, and attributing high-level ones,
such as beliefs. Assuming that learning of beliefs can take place
by observing own decision and beliefs estimation processes in
partially observable environments and using a simple feed-
forward deep learning model, we show that when learning
to predict others’ intentions and actions, faster and more
accurate predictions can be acquired if beliefs attribution is
learnt simultaneously with action and intentions prediction.
We show that the learning performance improves even when
observing agents with a different decision process and is
higher when observing beliefs-driven chunks of behaviour. We
propose that our architectural approach can be relevant for the
design of future adaptive social robots that should be able to
autonomously understand and assist human partners in novel
natural environments and tasks.
I. INTRODUCTION
Due to recent technological developments, the interactions
between AI and humans have become pervasive and hetero-
geneous, extending from voice assistants or recommender
systems supporting the online experience of millions of users
to autonomous driving cars. Principled models to represent
the human collaborators’ needs are being adopted [1] while
robotic perception in complex environments is becoming
more flexible and adaptive [2]–[5] even in social contexts,
robot sensory limits are starting to be actively managed [6],
[7]. However, robots and intelligent systems still have a
limited understanding of how sensory limits affect human
partners’ behaviour and lead them to rely on internal beliefs
about the state of the world. This strongly impacts human-
robot mutual understanding [8] and calls for an effort to
transfer the advance in robot perception management to
methods to better cope with human collaborators’ perceptual
limits [9]–[11].
The possibility of introducing in robots and AI systems
a Theory of Mind (TOM) [12], the ability to attribute to
*This work was not supported by any organization
1University of Essex, Colchester, UK
2Universit`
a degli Studi di Milano Bicocca, Milano, Italy
∗email:dimitri.ognibene@unimib.it
other agents’ beliefs, intentions, or mental states in general,
has recently raised hopes to further improve robots’ social
skills [13]–[16]. While some studies have explored human
partners’ tendency to attribute mental states to robots [17]–
[20], the expected practical impact of TOM led to a diverse
set of TOM implementations on robots. Several implementa-
tions relied on hardwired agents and task models that could
be applied to infer mental states in settings known at design
time [21]–[25]. A step forward is presented in [26] with an
algorithm to understand unknown agents relying upon Belief-
Desire-Intention models of previously met agents.
Recently, following [27] seminal work, several models
have introduced deep learning based TOM implementations
[28]–[33]. This novel approach, learning both beliefs and in-
tention attribution, should allow improved collaboration and
adaptive human-robot collaboration in complex environments
through a better understanding of humans’ mental states. In
this paper, we explore if the data-driven approach proposed
in [27] and related works leads to improved predictions
of the partner’s intentions, which is often the mental state
with the highest impact on the interaction performance.
The prediction of partners’ intentions, even within a system
producing prediction on several others’ unobservable mental
states, such as beliefs, will still rely only on the processing of
observable behavioural inputs, aka state-action trajectories.
In a purely supervised learning setting, such as that proposed
in [27], it is not immediate why performing an additional
set of predictions, increasing the demands on the social
perception system, should result in higher accuracy for the
prediction of others’ intentions. This approach introduces
additional complexity and noise that may hinder performance
(see [34]). Moreover, deep learning models as those proposed
in [27] are usually data hungry, which may further affect the
value of the approach. These factors may be some of the
reasons for the long time required for the full development
of TOM in infants [12], [35].
While all these considerations sound technically valid,
our results with simplified versions of the architecture pro-
posed in [27] show that the original hypothesis may be
true: learning is faster and more accurate if it takes place
simultaneously for the prediction of both intentions and
beliefs together. Our results also show that the impact of
learning beliefs attribution on intention prediction is stronger
in conditions of strong partial observability, e.g. when the
observed agent does not still know where his target is.
We found that when the system learns to predict intentions
and beliefs at the same time it can better disambiguate
and discard unrelated objects that are or have been in the
arXiv:2210.09435v1 [cs.RO] 17 Oct 2022