Optimizing American Sign Language Recognition with Binarized Neural Networks: A Comparative Study with Traditional Models

  • Shakeef Ahmed Rakin

    Department of Computer Science and Engineering, BRAC University
    66 Mohakhali, Dhaka 1212, Bangladesh

  • Md.Tahjid Ahsan

    Department of Computer Science and Engineering, BRAC University
    66 Mohakhali, Dhaka 1212, Bangladesh

  • Mohammed Intishar Rahman

    Department of Computer Science and Engineering, BRAC University
    66 Mohakhali, Dhaka 1212, Bangladesh

  • Afif Alamgir Lizel

    Department of Computer Science and Engineering, BRAC University
    66 Mohakhali, Dhaka 1212, Bangladesh

  • Md Sifat Mahmud

    Department of Computer Science and Engineering, BRAC University
    66 Mohakhali, Dhaka 1212, Bangladesh

  • Md Tanzim Reza

    Department of Computer Science and Engineering, BRAC University
    66 Mohakhali, Dhaka 1212, Bangladesh
  • BRAC University
PUBLISHEDBRAC University

BRAC University

Department of Computer Science and Engineering, BRAC University

Keywords


  • Artificial Intelligence (AI),
  • Sign Language,
  • Deep Learning,
  • American Sign Language (ASL),
  • Binarized Neural Networks (BNNs),
  • Larq

Abstract


Sign language is crucial for communication among individuals with hearing or speech impairments. Automated recognition systems are essential for learning and translating different sign language variants. However, these systems often face high computational demands and large memory footprints, limiting their use in real-time and resource-constrained environments. This research develops an optimized pipeline for American Sign Language (ASL) recognition, comparing Binarized Neural Networks (BNNs) with traditional full-precision neural networks. Using Larq, a library for training binarized models, we leverage BNNs' reduced memory and computational needs, suitable for embedded systems and edge devices. The study uses a dataset of ASL alphabet images, applying data augmentation to address data imbalance and occlusions. Both binarized and traditional models are trained and evaluated on accuracy, precision, recall, F1-score, memory footprints, and inference times. Results show that BNNs offer competitive performance with significantly lower computational requirements, demonstrating their potential for efficient and accessible ASL recognition systems.