Finally, we provide an empirical demonstration of evaluating a memory augmentation strategy for
GPT-2 [
13
] using human behavioral data and identify its limitations and strengths to inform future
research.
2 Considerations for biologically-inspired memory augmentation
2.1 Memory-augmentation from a cognitive lens
We argue that specifying appropriate linking hypotheses across domains will not only facilitate
novel biologically-inspired approaches, but will also provide a way to empirically evaluate different
hypotheses. In cognitive neuroscience, a linking hypothesis is a proposition for a formal mapping
between a neurobiological state and a psychological state [
14
,
15
], such as the firing of a single
neuron leading to a visual percept. A central aim of biologically-inspired AI is to formulate linking
hypotheses between a component in an AI system and a well-defined aspect of cognition. Strong
linking hypotheses should lead to a formal and quantifiable mapping between a representation in an
AI system and some neurobiological/psychological data, as has been demonstrated in some cases for
computer vision [
16
–
18
] and natural language [
19
–
22
]. These linking hypotheses must be specified at
the correct level of analysis [
15
,
23
], e.g., a modification of the equations to perform similarity search
on a database in a retrieval-augmented system should map to research on the biological mechanisms
of memory retrieval. In our view, proper accounts would be best derived through decomposing
the problem into computational subroutines appropriate for comparison across domains. Many AI
systems already assume a linking hypothesis between ANNs and human cognition without explicitly
stating them as hypotheses or evaluating them. Here, we briefly explore some of these hypotheses in
memory-augmented Transformers and propose possible mappings to findings in the human literature.
We divide memory-augmented Transformers into two general types. A static memory stores infor-
mation in a corpus of fixed size and content (e.g., a Wikipedia knowledge base), which it learns
to retrieve from during training [
24
,
25
,
9
]. The contents of a static memory do not get modified,
although they can be encoded in different formats such as raw text or embeddings [
25
,
26
]. Dynamic
memory mechanisms store new information as inputs that are being processed by the model. Training
the network involves learning both storage and retrieval policies. For example, new information may
be remembered or forgotten on the basis of input properties or model activations. Furthermore, inputs
may be transformed in some manner (e.g., through compression) before being stored in the external
memory [
7
]. Both static and dynamic memory-augmented Transformers have shown significant
improvements over non-augmented models when making predictions over long texts [7, 8, 27, 28].
These augmentation strategies do not map cleanly to the types of memory commonly delineated in
cognitive theories of human memory [
29
]. That said, classical memory taxonomies are often the
source of AI inspiration, with papers citing work on short- vs. long-term memory or episodic memory
[
30
,
31
,
10
,
11
]. In our view, a static memory could be like human semantic memory if it uses a
knowledge base, or it could be a fairly direct analog of episodic memory if it stores previously seen
examples [
24
]. Instead, our proposed division focuses on the subprocesses thought to be involved
in human long-term memory: encoding, consolidation, and retrieval [
32
]. Different strategies for
memory augmentation will therefore pursue different implementations of each subprocess, and can
draw direct inspiration from studies of that specific subprocess. Current work on memory-augmented
Transformers has already proposed separate mechanisms for each subprocess, although there is often
no direct link to human data. For example, there is a growing literature on retrieval augmentation
[
8
,
9
,
24
,
25
,
33
–
35
] that proposes similarity search as the retrieval mechanism. Other work has
proposed specific encoding policies which determine what to store and forget, either by exploiting
the attention weights [7] or learning which memories to forget [36].
2.2 Incorporating insights from human memory via policy modifications
Here we discuss some findings from the human memory literature to demonstrate how they may be
used to inform policy modifications in memory-augmented Transformers. Lexical properties (e.g.,
written-frequency, word length, animacy, etc.) serve as strong predictors of subsequent memory for
individual words and lists [
37
–
39
]. Furthermore, humans have been shown to have the remarkable
ability to remember whether they have seen an image from up to 10,000 images after only a single
exposure [
40
]. The properties that determine the memorability of an image are thought to be
multifaceted, including high-level properties such as emotional valence [
41
] and overall semantic
2