• SIGN LANGUAGE RECOGNITION USING CONVOLUTIONAL NEURAL NETWORKS

    Author(s):
    Prabhakara Uyyala (see profile)
    Date:
    2022
    Item Type:
    Article
    Permanent URL:
    https://doi.org/10.17613/47ga-zw60
    Abstract:
    Sign Language Recognition (SLR) targets on interpreting the sign language into text or speech, so as to facilitate the communication between deaf-mute people and ordinary people. This task has broad social impact, but is still very challenging due to the complexity and large variations in hand actions. Existing methods for SLR use hand-crafted features to describe sign language motion and build classification models based on those features. However, it is difficult to design reliable features to adapt to the large variations of hand gestures. To approach this problem, we propose a novel convolution neural network (CNN) which extracts discriminative spatial-temporal features from raw video stream automatically without any prior knowledge, avoiding designing features. To boost the performance, multi-channels of video streams, including color information, depth clue, and body joint positions, are used as input to the CNN in order to integrate color, depth and trajectory information. We validate the proposed model on a real dataset collected with Microsoft Kinect and demonstrate its effectiveness over the traditional approaches based on hand-crafted features.
    Metadata:
    Published as:
    Journal article    
    Status:
    Published
    Last Updated:
    5 months ago
    License:
    Attribution-NonCommercial
    Share this:

    Downloads

    Item Name: pdf 125-jicr-january-2022.pdf
      Download View in browser
    Activity: Downloads: 108