Publication

An HMM Approach with Inherent Model Selection for Sign Language and Gesture Recognition

Oya Aran, Sandrine Tornay
2020
Conference paper
Abstract

HMMs have been the one of the first models to be applied for sign recognition and have become the baseline models due to their success in modeling sequential and multivariate data. Despite the extensive use of HMMs for sign recognition, determining the HMM structure has still remained as a challenge, especially when the number of signs to be modeled is high. In this work, we present a continuous HMM framework for modeling and recognizing isolated signs, which inherently performs model selection to optimize the number of states for each sign separately during recognition. Our experiments on three different datasets, namely, German sign language DGS dataset, Turkish sign language HospiSign dataset and Chalearn14 dataset show that the proposed approach achieves better sign language or gesture recognition systems in comparison to the approach of selecting or presetting the number of HMM states based on k-means, and yields systems that perform competitive to the case where the number of states are determined based on the test set performance.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Related concepts (21)
Sign language
Sign languages (also known as signed languages) are languages that use the visual-manual modality to convey meaning, instead of spoken words. Sign languages are expressed through manual articulation in combination with non-manual markers. Sign languages are full-fledged natural languages with their own grammar and lexicon. Sign languages are not universal and are usually not mutually intelligible, although there are also similarities among different sign languages.
Gesture recognition
Gesture recognition is a topic in computer science and language technology with the goal of interpreting human gestures via mathematical algorithms. It is a subdiscipline of computer vision. Gestures can originate from any bodily motion or state, but commonly originate from the face or hand. Focuses in the field include emotion recognition from face and hand gesture recognition since they are all expressions. Users can make simple gestures to control or interact with devices without physically touching them.
British Sign Language
British Sign Language (BSL) is a sign language used in the United Kingdom and is the first or preferred language among the deaf community in the UK. Based on the percentage of people who reported 'using British Sign Language at home' on the 2011 Scottish Census, the British Deaf Association estimates there are 151,000 BSL users in the UK, of whom 87,000 are Deaf. By contrast, in the 2011 England and Wales Census 15,000 people living in England and Wales reported themselves using BSL as their main language.
Show more
Related publications (33)

On the Recognition Performance of BioHashing on state-of-the-art Face Recognition models

Sébastien Marcel, Hatef Otroshi Shahreza

Face recognition has become a popular authentication tool in recent years. Modern state-of-the-art (SOTA) face recognition methods rely on deep neural networks, which extract discriminative features from face images. Although these methods have high recogn ...
IEEE2021

Data-Driven Movement Subunit Extraction from Skeleton Information for Modeling Signs and Gestures

Marzieh Razavi, Sandrine Tornay

Sequence modeling for signs and gestures is an open research problem. In thatdirection, there is a sustained effort towards modeling signs and gestures as a se-quence of subunits. In this paper, we develop a novel approach to infer movementsubunits in a da ...
Idiap2019

Tree-structured Classifier for Acceleration-based Activity and Gesture Recognition on Smartwatches

Karl Aberer, Julien Eberle, Dipanjan Chakraborty

This paper proposes a new method for recognizing both activities and gestures by using acceleration data collected on a smartwatch. While both activity recognition techniques and gesture recognition techniques employ acceleration data, these techniques are ...
Ieee2016
Show more

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.