Recognizing touch gestures for social human-robot interaction
[ X ]
Tarih
2015
Yazarlar
Dergi Başlığı
Dergi ISSN
Cilt Başlığı
Yayıncı
Assoc Computing Machinery
Erişim Hakkı
info:eu-repo/semantics/closedAccess
Özet
In this study, we performed touch gesture recognition on two sets of data provided by "Recognition of Social Touch Gestures Challenge 2015". For the first dataset, dubbed Corpus of Social Touch (CoST), touch is performed on a mannequin arm, whereas for the second dataset (Human-Animal Affective Robot Touch HAART) touch is performed in a human-pet interaction setting. CoST includes 14 gestures and HAART includes 7 gestures. We used the pressure data, image features, Hurst exponent, Hjorth parameters and autoregressive model coefficients as features, and performed feature selection using sequential forward floating search. We obtained classification results around 60%-70% for the HAART dataset. For the CoST dataset, the results range from 26% to 95% depending on the selection of the training/test sets.
Açıklama
2015 ACM International Conference on Multimodal Interaction -- NOV 09-13, 2015 -- Seattle, WA
Altun, Kerem/0000-0002-5493-8921
Altun, Kerem/0000-0002-5493-8921
Anahtar Kelimeler
Gesture Recognition, Human-Robot Interaction, Random Forests, Feature Selection, Sequential Floating Forward Search
Kaynak
Icmi'15: Proceedings of the 2015 Acm International Conference on Multimodal Interaction
WoS Q Değeri
N/A
Scopus Q Değeri
N/A