Skip to content

atalaydenknalbant/ASL-Detector-yolo

Repository files navigation

License: AGPL v3 visitors HF

Sign Language Translator

Overview

The aim of this project was to develop an application that can translate American Sign Language (ASL) into written language, using object detection techniques and multiple data sources.​

Run Inference On Hugging Face

https://huggingface.co/spaces/atalaydenknalbant/ASL-Detector-YOLO

Trained Models Link

https://huggingface.co/atalaydenknalbant/asl-models/tree/main

Defining The Problem ​And The Solution

  • We want to improve communication and inclusion between people who use ASL and those who do not.​
  • Our project uses computer vision and machine learning to detect and translate ASL signs into words.​
  • Our project is an example of how technology can help overcome linguistic and cultural differences.

SDGS

  1. SDG 4.5 mission is: "Achieve gender equality and inclusive education for all by 2030."​
  • Our app helps to achieve SDG 4.5 by enabling deaf and hard-of-hearing students to access quality education and communicate with their peers and teachers.
  1. SDG 10.3 mission is: "Promote equal opportunity and fight discrimination through laws and policies."​
  • Our app helps to achieve SDG 10.3 by empowering deaf and hard-of-hearing people to overcome communication barriers and access equal opportunities in society.

Our Framework

DATASET

  • Available on Roboflow Universe:​ 6879 total image. 6455 train – 276 val – 148 test.​
  • 26 Class: A, B, C, D, E, F, G, H, I, J, K, L, M, N, O, P, Q, R, S, T, U, V, W, X, Y, Z ​
  • 3 Different Datasets​
  • Images dimension: ​ (384, 384), (416,416) (640,480)

Model

Our model is the Yolov9-small model, which was pretrained on the COCO dataset for object detection.

Results

We used various metrics and methods to evaluate our model’s performance and accuracy. On the left We used a confusion matrix, which is a table that shows how well our model predicted the correct classes for each image. On the right We also used graphs to visualize the training and validation loss and accuracy, as well as the precision and recall metrics, over epochs. These are some of the results we achieved. We will look closer to metrics in the next slide.

Metrics

We used three metrics: box precision, recall, mAP50. These metrics are commonly used for object detection tasks, such as detecting and translating ASL signs. We calculated these metrics for each class, corresponding to each letter of the alphabet, and also for the overall model. We achieved high values for all these metrics, as shown by the table on the left. We are very satisfied with our model’s results, and we believe that our model can translate ASL signs into words effectively and reliably.

Test Results

Labels

Predicts

Here are the results of labels we predicted by using our model. As you can see our model didn't miss any letter at given dataset.

Demo

Summary

Our project has the following aspects:

License

Project licensed under AGPL-3.0. Also project contains Ultralytics software licensed under AGPL-3.0.

Citation

Releases

No releases published

Packages

 
 
 

Contributors