For cycle-based segmentation, a distance-based metric
is usually applied to establish a match threshold for
classifying each cycle as genuine, or an impostor [5,7].
Recently researchers have focused on the extraction of
statistical and frequency domain features from the cycles
and the application of machine learning algorithms. On
the other hand, researchers have established a machine
learning pipeline to train the authentication models for the
fixed-length segmentation case. [3,6].
Authentication models trained using the fixed-length
frame with a machine learning pipeline have usually
outperformed cycle-based approaches with distance
metrics, e.g., Euclidean or Manhattan distances or distance
measures such as Dynamic Time Warping (DTW) [8,9,10].
An experiment on the same dataset showed a difference of
13.6% [10] with the frame-based training achieving 92.7%
accuracy and cycle-based approach achieving 79.1%.
Recent studies with a better machine learning pipeline have
achieved error rates under 5% [3].
Because authentication systems play an essential role
in securing users’ data and privacy, it is paramount
that the authentication systems can defend themselves
from adversarial attacks. Zhao et al. [11] examined the
vulnerability of ABGait from random-vector attacks. The
performance of ABGait is generally evaluated in terms
of the number of false accepts (or false positives) and
false rejects (false negatives). The lower these numbers
are, the better the ABGait is. However, Zhao et al. [11]
pointed out that the false-positive rates of ABGait could
be misleading as they are computed entirely based on the
impostor samples that are available at the time of testing.
Interestingly, until Zhao et al. [11], researchers had
overlooked the possibility of someone attacking ABGait
with random feature vectors with the assumption that the
attacker would know the length of the feature vectors that
are used to train the ABGait. Zhao et al. [11] showed that
ABGait accepts even uniform random inputs. Additionally,
they found that the probability of random vector acceptance
is much higher than the false-positive rates. They suggested
that the acceptance and rejection regions created during the
ABGait training are much bigger than the false accept and
false reject rates. Therefore the performance evaluation
of ABGait should include both measures, i.e., False
Acceptance Rates (FAR) and Acceptance Region (AR), to
assess the ability of ABGait to defend itself from active
adversarial attempts such as random-vector attacks [11].
1.3. Adversarial Scenarios
Its commonly believed that behavioral biometrics such
as ABGait require more effort to circumvent than physical
biometrics. However, some studies have pointed out that
ABGait is vulnerable to treadmill-assisted imitation attacks
[9] and random-vector attacks [11]. These two attack
paradigms are substantially different. The former requires
minimal system knowledge, is difficult to launch, and
requires artifacts such as a treadmill. In contrast, the latter
merely requires the generation of random vectors. However,
it also assumes that the attacker would have access to the
authentication API and the knowledge of feature space
(length of feature vectors, normalization methods, etc.)
[11]. Considering the feasibility and ease with which a
random vector attack can be executed, we focus on the same
in this study.
1.4. Possible Countermeasures
The possible countermeasures to the random-vector
attack include training the authentication models using
synthetic data. Zhao et al. [11] used synthetic
noise generated around genuine samples to train the
authentication models. The generated noise around the
genuine samples was labeled as impostors during the
training process to increase the impostor samples’ variance.
The model showed greater resilience to random-vector
attacks than the models without synthetic data-based
training. Quantitatively, the models achieved a reduced
area of the acceptance region, representing a potentially
errant input as genuine. The idea worked on the
dataset [12] that was studied in [11] but needs to be
tested on multiple datasets, including ones collected from
smartwatches. Inspired by the previous countermeasure, we
propose using Conditional Tabular Generative Adversarial
Networks (CTGAN) [13] on impostor samples to increase
the variance of the impostor samples and test its usefulness
in mitigating the random-vector attack while maintaining
the reported performance. GAN has been successfully used
to safeguard gait-based key generation from vision-based
side-channel attacks in the past [14].
1.5. Main Contributions
First, we implemented vanilla ABGait (vABGait) with
no mitigation technique in the pipeline. Then we tested the
same on three different datasets consisting of a different
number of users, samples per user, and feature set and
computed the performance measures, i.e., False Accept
Rate (FAR), False Reject Rate (FRR), Acceptance Region
(AR), and Half Total Error Rate (HTER) under the
zero-effort and random-vector attack scenarios. Second, we
included the mitigation technique proposed in Zhao et al.
[11] which resulted in (βABGait), and evaluated the same
on all three datasets using the aforementioned metrics under
the zero-effort and random-vector attack scenarios. Third,
we introduced iCTGAN, a CTGAN-assisted impostor
samples generator, in the ABGait training pipeline,
evaluated its performance on the three datasets, and
compared its performance with βABGait and vABGait.
The pre-processed dataset and code is available at [15].