Cross-Modal Generation of Tactile Friction Coefficient From Audio and Visual Measurements by Transformer.

IEEE Trans. Instrum. Meas.(2023)

引用 1|浏览0
暂无评分
摘要
Generating tactile data (e.g., friction coefficient) from audio and visual modalities can avoid time-consuming practical measurements and ensure high-fidelity haptic rendering of surface textures. In this article, we present a Transformer-based method for cross-modal generation of the tactile friction coefficient. Using the self-attention mechanism, we jointly encode the amplitude and phase information of audio spectrums and RGB images to extract global and local features. Then, we convert the joint coding features into tactile decoding features using a Transformer module in a bottleneck converter. We continuously decode and reconstruct them to obtain amplitude and phase information of tactile friction coefficients. Finally, we convert this information into 1-D friction coefficients using inverse short-time Fourier transform (ISTFT). Evaluations of the LMT Haptic Material Database confirm the obvious performance improvement of the proposed method. Furthermore, with the generated friction by the Transformer and the custom-designed electrovibration device, a novel rendering method is proposed, which simultaneously utilizes the amplitude and frequency of driving signals to display tactile textures on touchscreens. User experiments are organized to evaluate the rendering fidelity of the generated friction coefficient. The two-way analysis of variance (ANOVA) with repeated measurements indicates that the rendering fidelity of the Transformer method is significantly improved compared with the contrast methods ( p < 0.05).
更多
查看译文
关键词
Audio data,electrovibration,surface haptics,tactile friction coefficient,Transformer,visual data
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要