Visual-Tactile Cross-Modal Data Generation using Residue-Fusion GAN with Feature-Matching and Perceptual Losses
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Pages (from-to) | 7525-7532 |
Number of pages | 8 |
Journal / Publication | IEEE Robotics and Automation Letters |
Volume | 6 |
Issue number | 4 |
Online published | 9 Jul 2021 |
Publication status | Published - Oct 2021 |
Link(s)
Abstract
Existing psychophysical studies have revealed that the cross-modal visual-tactile perception is common for humans performing daily activities. However, it is still challenging to build the algorithmic mapping from one modality space to another, namely the cross-modal visual-tactile data translation/generation, which could be potentially important for robotic operation. In this paper, we propose a deep-learning-based approach for cross-modal visual-tactile data generation by leveraging the framework of the generative adversarial networks (GANs). Our approach takes the visual image of a material surface as the visual data, and the accelerometer signal induced by the pen-sliding movement on the surface as the tactile data. We adopt the conditional-GAN (cGAN) structure together with the residue-fusion (RF) module, and train the model with the additional feature-matching (FM) and perceptual losses to achieve the cross-modal data generation. The experimental results show that the inclusion of the RF module, and the FM and the perceptual losses significantly improves cross-modal data generation performance in terms of the classification accuracy upon the generated data and the visual similarity between the ground-truth and the generated data.
Research Area(s)
- generative adversarial networks (GANs), Deep Learning, Cross-modal perception, Visual-tactile, Robot sensing systems, visual perception
Citation Format(s)
Visual-Tactile Cross-Modal Data Generation using Residue-Fusion GAN with Feature-Matching and Perceptual Losses. / Cai, Shaoyu; Zhu, Kening; Ban, Yuki et al.
In: IEEE Robotics and Automation Letters, Vol. 6, No. 4, 10.2021, p. 7525-7532.
In: IEEE Robotics and Automation Letters, Vol. 6, No. 4, 10.2021, p. 7525-7532.
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review