This project is a real-time hand gesture recognition system that uses computer vision and deep learning technologies to classify hand gestures from webcam input. The system leverages MediaPipe for hand landmark detection and a custom Convolutional Neural Network (CNN) for gesture classification.
- Real-time hand gesture recognition
- Uses MediaPipe for hand landmark detection
- Custom CNN model for gesture classification
- Developed and trained on Google Colab
- Supports multiple gesture categories
- Direct Google Drive file access
- Compressed image dataset handling
- Automated model training and checkpointing
- GPU/TPU acceleration for faster computations
- Development Platform: Google Colab
- Hardware Acceleration: TPU
- Computer Vision: OpenCV (cv2)
- Hand Tracking: MediaPipe
- Deep Learning: TensorFlow/Keras
- Programming Language: Python
- Captures hand landmark images using webcam
- Processes and saves landmark images for training
- Supports different hand configurations (left/right, normal/flipped)
- Prepares and preprocesses image dataset
- Builds a Convolutional Neural Network (CNN)
- Trains and validates the gesture recognition model
- Saves the best performing model
- Loads pre-trained model
- Processes real-time webcam input
- Performs hand gesture recognition
- Displays prediction results
- Google Account
- Google Colab access
- Prepared image dataset
- Open Google Colab
- Create new notebook
- Upload or link to required Python scripts
- Mount Google Drive
- Upload compressed image dataset
- Run training notebook (Project_HGR.ipynb)
After training in Colab:
- Download the best performing model
- Use
live_cam_test.py
for real-time gesture recognition - Ensure all dependencies are installed locally
- Increase training dataset diversity
- Implement data augmentation
- Experiment with model architectures
- Add more gesture categories
- Requires good lighting conditions
- Performance depends on training data quality
- Currently supports a limited number of gesture categories