This project focuses on developing a comprehensive Arabic Sign Language (ASL) Recognition System using advanced deep learning techniques. Our goal is to bridge communication barriers for the deaf and hard-of-hearing community by creating an accurate and efficient sign language recognition system.
- Develop an Arabic Sign Language recognition system using state-of-the-art deep learning models
- Compare effectiveness of different neural network architectures:
- Convolutional Neural Networks (CNNs)
- Long Short-Term Memory (LSTM) Networks
- Graph Neural Networks (GNNs)
- Transfer Learning with pre-trained models
- Evaluate models based on accuracy, efficiency, and practical usability
- Create a robust data preprocessing pipeline for sign language recognition
- TensorFlow/Keras - Primary deep learning framework
- YOLO - Object detection for hand recognition
- ResNet101V2 - Transfer learning backbone
- OpenCV - Computer vision and image processing
- NumPy - Numerical computations
- Pandas - Data manipulation and analysis
- Matplotlib - Data visualization and plotting
- Jupyter Notebook - Development environment
- Python - Primary development language
vip/
├── README.md # Project documentation
├── Code/
│ ├── code.ipynb # Main implementation notebook
│ ├── best.pt # Trained YOLO model (14MB)
│ ├── logo.png # Project logo
│ ├── Input/ # Input data directory
│ └── Output/ # Processed output directory
├── Final Report.pdf # Comprehensive project report
└── [Presentation File] # Project presentation
Raw Sign Language Data → YOLO Hand Detection → Data Cleaning → Preprocessing → Model Training → Evaluation
- Hand Detection: YOLO model identifies and extracts hand regions
- Data Cleaning: Removes irrelevant background and noise
- Class Organization: Sorts images into class-specific folders
- Data Augmentation: Enhances dataset diversity
- Exploratory Data Analysis (EDA): Understanding data distribution and characteristics
- Feature Extraction: Extracting relevant features for sign recognition
- Model Architecture Design: Implementing various neural network approaches
- Hyperparameter Tuning: Optimizing model performance
- Accuracy - Overall model performance
- Precision - Correct positive predictions
- Recall - Ability to find all positive instances
- F1-Score - Harmonic mean of precision and recall
Python 3.7+
Jupyter Notebook
CUDA-compatible GPU (recommended)
- Clone the repository:
git clone [repository-url]
cd vip
- Install required dependencies:
pip install tensorflow opencv-python numpy pandas matplotlib
pip install torch torchvision # For YOLO model
pip install jupyter notebook
- Navigate to the code directory:
cd Code
- Launch Jupyter Notebook:
jupyter notebook
-
Open the main implementation:
- Open
code.ipynb
in Jupyter Notebook - Follow the step-by-step implementation
- Open
-
Data Preparation:
- Place input sign language images in the
Input/
directory - Run the data preprocessing cells in the notebook
- Place input sign language images in the
-
Model Training:
- Execute the model training sections
- Monitor training progress and metrics
-
Evaluation:
- Run evaluation cells to assess model performance
- View results and comparative analysis
- CNN Models: Specialized for image feature extraction
- LSTM Networks: Capturing temporal dependencies in sign sequences
- Graph Neural Networks: Modeling hand joint relationships
- Transfer Learning: Leveraging pre-trained ResNet101V2
- YOLO Integration: Accurate hand detection and extraction
- Automated Cleaning: Removes background noise and irrelevant data
- Smart Organization: Class-based data structuring
- Multi-metric Assessment: Accuracy, precision, recall, F1-score
- Comparative Analysis: Side-by-side model performance evaluation
- Efficiency Metrics: Training time and inference speed analysis
- High-Accuracy Model: Achieving optimal recognition rates for Arabic sign language
- Efficiency Optimization: Balancing accuracy with computational efficiency
- Practical Implementation: Ready-to-deploy sign language recognition system
- Research Contribution: Advancing Arabic sign language recognition research
Note: Detailed results and performance metrics are available in the
Final Report.pdf
and within the Jupyter notebook implementation.
Key achievements:
- ✅ Successful implementation of multiple deep learning architectures
- ✅ Effective YOLO-based hand detection pipeline
- ✅ Comprehensive comparative analysis of model performance
- ✅ Robust data preprocessing and augmentation pipeline
- Real-time Recognition: Implementing live video sign language recognition
- Mobile Application: Developing mobile app for accessibility
- Extended Vocabulary: Expanding to larger Arabic sign language vocabulary
- Multi-language Support: Adding support for other sign languages
- Edge Deployment: Optimizing models for edge device deployment
- 📄 Final Report: Comprehensive technical documentation in
Final Report.pdf
- 💻 Code Documentation: Detailed comments and explanations in
code.ipynb
- 🎯 Methodology: Step-by-step implementation guide in the notebook
This project is part of an academic VIP (Vertically Integrated Projects) program. For collaboration or questions:
- Academic Institution: Multimedia University (MMU)
- Project Type: Visual Information Processing
For questions or collaboration opportunities, please contact the team members through MMU academic channels.
- Multimedia University (MMU) - Academic support and resources
- VIP Program - Providing the platform for this research
- Open Source Community - For the excellent tools and frameworks
- Arabic Sign Language Community - For the inspiration and importance of this work