Interpreting sign language images into Text-Based using YOLOv7

Penulis

  • Julius Sabilala College of Science and Technology, Guimaras State University-Mosqueda Campus, Alaguisoc, Jordan, Guimaras, 5044, Philippines
  • Rv Jan Zaldo Bautista College of Science and Technology, Guimaras State University-Mosqueda Campus, Alaguisoc, Jordan, Guimaras, 5044, Philippines
  • Lea Mae Gambol Tortogo College of Science and Technology, Guimaras State University-Mosqueda Campus, Alaguisoc, Jordan, Guimaras, 5044, Philippines
  • Lerry Joy Ga Juntarciego College of Science and Technology, Guimaras State University-Mosqueda Campus, Alaguisoc, Jordan, Guimaras, 5044, Philippines
  • Charles Vincent Gayoma College of Science and Technology, Guimaras State University-Mosqueda Campus, Alaguisoc, Jordan, Guimaras, 5044, Philippines
  • Adrian Forca College of Science and Technology, Guimaras State University-Mosqueda Campus, Alaguisoc, Jordan, Guimaras, 5044, Philippines

Abstrak

Sign Language recognition remains a significant challenge due to variations in hand gestures, occlusions, and environmental factors. This study introduces a YOLOv7-based sign language recognition system designed to interpret American Sign Language (ASL) images into text in real time, enhancing communication accessibility for individuals with hearing and speech impairments. The primary objectives are to develop an accurate detection model, improve real-time gesture interpretation, and optimize system performance for accessibility. The study follows the CRISP-DM (Cross-Industry Standard Process for Data Mining) methodology, including data collection, preprocessing, model training, evaluation, and deployment. The system was developed using YOLOv7, Python, and OpenCV, with an ASL dataset sourced from Roboflow and formatted in YOLOv7 PyTorch. Image preprocessing techniques such as normalization, resizing, and data augmentation were applied to enhance detection accuracy. The application integrates a real-time gesture recognition system, where detected ASL signs are instantly translated into text. Results demonstrate high detection accuracy for most ASL letters, but certain gestures such as J, Z, S, and T posed challenges due to similar hand shapes and motion-based characteristics. Optimization efforts included dataset expansion, refined annotations, and hyperparameter tuning to improve model precision. The system significantly enhances real-time ASL recognition, offering a scalable, AI-powered assistive tool for the deaf and hard of hearing community. This research contributes to machine learning-driven accessibility solutions, bridging the communication gap between sign language users and non-signers.

Diterbitkan

2025-07-31

Cara Mengutip

Sabilala, J., Bautista, R. J. Z. ., Tortogo, L. M. G. ., Juntarciego, . L. J. G., Gayoma, C. V. ., & Forca, A. (2025). Interpreting sign language images into Text-Based using YOLOv7. Komputasi: Jurnal Ilmiah Ilmu Komputer Dan Matematika, 22(2), 90–98. Diambil dari https://komputasi-fmipa.unpak.ac.id/index.php/komputasi/article/view/13