American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach

Sign language is intentionally designed to allow deaf and dumb communities to convey messages and to connect with society. Unfortunately, learning and practicing sign language is not common among society; hence, this study developed a sign language recognition prototype using the Leap Motion Control...

Full description

Bibliographic Details
Main Authors: Teak-Wei Chong, Boon-Giin Lee
Format: Article
Language:English
Published: MDPI AG 2018-10-01
Series:Sensors
Subjects:
Online Access:http://www.mdpi.com/1424-8220/18/10/3554
id doaj-b414e160296842f584ca656d3789ebef
record_format Article
spelling doaj-b414e160296842f584ca656d3789ebef2020-11-25T00:44:11ZengMDPI AGSensors1424-82202018-10-011810355410.3390/s18103554s18103554American Sign Language Recognition Using Leap Motion Controller with Machine Learning ApproachTeak-Wei Chong0Boon-Giin Lee1Department of Electronics Engineering, Keimyung University, Daegu 42601, KoreaDepartment of Electronics Engineering, Keimyung University, Daegu 42601, KoreaSign language is intentionally designed to allow deaf and dumb communities to convey messages and to connect with society. Unfortunately, learning and practicing sign language is not common among society; hence, this study developed a sign language recognition prototype using the Leap Motion Controller (LMC). Many existing studies have proposed methods for incomplete sign language recognition, whereas this study aimed for full American Sign Language (ASL) recognition, which consists of 26 letters and 10 digits. Most of the ASL letters are static (no movement), but certain ASL letters are dynamic (they require certain movements). Thus, this study also aimed to extract features from finger and hand motions to differentiate between the static and dynamic gestures. The experimental results revealed that the sign language recognition rates for the 26 letters using a support vector machine (SVM) and a deep neural network (DNN) are 80.30% and 93.81%, respectively. Meanwhile, the recognition rates for a combination of 26 letters and 10 digits are slightly lower, approximately 72.79% for the SVM and 88.79% for the DNN. As a result, the sign language recognition system has great potential for reducing the gap between deaf and dumb communities and others. The proposed prototype could also serve as an interpreter for the deaf and dumb in everyday life in service sectors, such as at the bank or post office.http://www.mdpi.com/1424-8220/18/10/3554human-computer interactionmachine learningsign language recognitionLeap Motion Controllersupport vector machinedeep neural networkmulti-class classificationAmerican Sign Language
collection DOAJ
language English
format Article
sources DOAJ
author Teak-Wei Chong
Boon-Giin Lee
spellingShingle Teak-Wei Chong
Boon-Giin Lee
American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach
Sensors
human-computer interaction
machine learning
sign language recognition
Leap Motion Controller
support vector machine
deep neural network
multi-class classification
American Sign Language
author_facet Teak-Wei Chong
Boon-Giin Lee
author_sort Teak-Wei Chong
title American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach
title_short American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach
title_full American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach
title_fullStr American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach
title_full_unstemmed American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach
title_sort american sign language recognition using leap motion controller with machine learning approach
publisher MDPI AG
series Sensors
issn 1424-8220
publishDate 2018-10-01
description Sign language is intentionally designed to allow deaf and dumb communities to convey messages and to connect with society. Unfortunately, learning and practicing sign language is not common among society; hence, this study developed a sign language recognition prototype using the Leap Motion Controller (LMC). Many existing studies have proposed methods for incomplete sign language recognition, whereas this study aimed for full American Sign Language (ASL) recognition, which consists of 26 letters and 10 digits. Most of the ASL letters are static (no movement), but certain ASL letters are dynamic (they require certain movements). Thus, this study also aimed to extract features from finger and hand motions to differentiate between the static and dynamic gestures. The experimental results revealed that the sign language recognition rates for the 26 letters using a support vector machine (SVM) and a deep neural network (DNN) are 80.30% and 93.81%, respectively. Meanwhile, the recognition rates for a combination of 26 letters and 10 digits are slightly lower, approximately 72.79% for the SVM and 88.79% for the DNN. As a result, the sign language recognition system has great potential for reducing the gap between deaf and dumb communities and others. The proposed prototype could also serve as an interpreter for the deaf and dumb in everyday life in service sectors, such as at the bank or post office.
topic human-computer interaction
machine learning
sign language recognition
Leap Motion Controller
support vector machine
deep neural network
multi-class classification
American Sign Language
url http://www.mdpi.com/1424-8220/18/10/3554
work_keys_str_mv AT teakweichong americansignlanguagerecognitionusingleapmotioncontrollerwithmachinelearningapproach
AT boongiinlee americansignlanguagerecognitionusingleapmotioncontrollerwithmachinelearningapproach
_version_ 1725275804069265408