organizes in 2013 a challenge and workshop on multi-modal gesture
recognition from 2D and 3D video data using Kinect, in conjunction
with ICMI 2013, December 9-13, Sidney, Australia.
revolutionizing the field of gesture recognition given the set of
input data modalities it provides, including RGB image, depth image
(using an infrared sensor), and audio. Gesture recognition is
genuinely important in many multi-modal interaction and computer
vision applications, including image/video indexing, video
surveillance, computer interfaces, and gaming. It also provides
excellent benchmarks for algorithms. The recognition of continuous,
natural signing is very challenging due to the multimodal nature of
the visual cues (e.g., movements of fingers and lips, facial
expressions, body pose), as well as technical limitations such as
spatial and temporal resolution and unreliable depth cues.
The Multi-modal Challenge workshop will be devoted to the presentation of most recent and challenging techniques from multi-modal gesture recognition. The committee encourages paper submissions in the following topics (but not limited to):
-Multi-modal descriptors for gesture recognition
-Fusion strategies for gesture recognition
-Multi-modal learning for gesture recognition
-Data sets and evaluation protocols for multi-modal gesture recognition
-Applications of multi-modal gesture recognition
The results of the challenge will be discussed at the workshop. It features a quantitative evaluation of automatic gesture recognition from a multi-modal dataset recorded with Kinect (providing RGB images of face and body, depth images of face and body, skeleton information, joint orientation and audio sources), including about 14,000 Italian gestures from several users. The emphasis of this edition of the competition will be on multi-modal automatic learning of a vocabulary of 20 types of Italian gestures performed by several different users while explaining a history, with the aim of performing user independent continuous gesture recognition combined with audio information.
Additionally, the challenge includes a live competition of demos/systems of applications based on multi-modal gesture recognition techniques. Demos using data from different modalities and different kind of devices are welcome. The demos will be evaluated in terms of multi-modality, technical quality, and applicability.
workshop papers and top three ranked participants of the quantitative evaluation will be invited to present their work at ICMI 2013 and their papers will be published in the ACM proceedings. Additionally,
there will be travel grants (based on availability) and the possibility to be invited to present extended versions of their works to a special issue in a high impact factor journal. Moreover, all
three top ranking participants in both, quantitative and qualitative challenges will be awarded with a ChaLearn winner certificate and an economic prize (based on availability). We will also announce a best
paper and best student paper awards among the workshop contributions.
April 30: Initial sample data for the 2013 Multi-modal Gesture Recognition Challenge released. The final development data will be released on May 22th.
April 1, 2013: Initial sample data for the 2013 Multi-modal Gesture Recognition Challenge will be released on April 30.
December 15, 2012: The 2011/2012 one-shot-learning challenges ended
ROUND 2 RESULTS:
ROUND 1 RESULTS: