method of inducing rotation for in-hand manipulation using gravity. Such manipulation, known as
pivoting, is key to performing tasks requiring an object to be at a specific relative angle to a gripper,
such as stacking shelves [5].
This paper proposes a method for a robot with a parallel gripper to rotate a long object grasped away
from its center of mass to a desired final relative orientation. This aims to allow parallel grippers to
robustly reorient objects into a desired orientation without having to regrasp the object. To achieve
this task, we focus on addressing two challenges: tracking the position of the object and controlling
the gripper to allow for gravitational pivoting towards the target angle.
Vision-based methods to track the object often make use of an eye-in-hand camera. However, the
gripper will often occlude the object, making it difficult to estimate the angle of the object accu-
rately [6]. An alternative is to use an externally placed camera. However, this necessitates the robot
moving to a fixed position in front of the camera for each manipulation. We use purely tactile infor-
mation to track the object to avoid these issues. We design an LSTM-based neural network model,
RSE-LSTM, which uses tactile information to predict a held objects’ relative angular position and
angular velocity.
Previous approaches to controlling the gripper often used model-based approaches, which required
information about the object, such as shape, mass, and friction. In contrast, we design a simple
gripper controller that assumes no a priori knowledge about the object parameters to allow for gen-
eralization to unseen objects.
We collect a real-world force-based tactile dataset, on ten household objects. This dataset is anno-
tated with both angular position and velocity measurements. RSE-LSTM is trained on this dataset,
and the results are reported with respect to both unseen data and unseen objects. We further validate
our approach experimentally on unseen objects.
The contributions of our paper are threefold:
• An annotated dataset containing gravitational pivoting with 10 household objects.
• A LSTM-based neural network which can predict both the velocity and angle of an object using
only tactile information.
• A grip controller, which can adjust the width of the gripper to allow an object to pivot in-hand to
achieve a required relative angle.
2 Related Works
Slip measurement. Slip detection is often framed as a binary problem, with machine learning
models predicting either slip or no slip. This is achieved with the use of visual sensors [7], force-
based sensors [8]. or optical sensors [9]. Various machine learning techniques have been used
including: Support Vector Machines [10,9], MLPs [11] and LSTM models [12]. Alternatively,
Convolutional Neural Networks (CNN) have been used to both detect and classify the type of slip
as either translational or rotational [13]. LSTM models have been used to determine the direction
of rotational slip [14], or the overall direction of the combination of rotational and translational
slip [15].
The domain of quantitative slip measurement is comparatively underexplored. Previous works mea-
sure the amount of translational slip using image-based tactile sensors [16]. Alternatively, visual
gel-based tactile sensors have been used to measure the rotation angle using a model-based ap-
proach [17]. However, to our knowledge, the use of force-based tactile sensing has not been ex-
plored, which is the focus of this paper.
Induced rotation. To induce rotation in a held object, previous work has made use of the external
environment to apply a torque or force on the held object [18,19,20]. However, rotation can also
be induced without any interaction with external objects. For example, the robot can perform a
swinging motion using the end-effector, where the velocity of the swing aims to bring an object to a
desired angle [21,22,23].
Alternatively, by loosening the grip on the object, gravity can be used to induce a rotation in the
object [24,25,26,27,28,29]. These approaches are model-based and rely on prior knowledge of
important parameters of the system, such as shape, mass and friction of the held object. Our work
assumes no prior knowledge about such parameters to allow for generalization to unseen objects.
2