IEEE World AI IoT Congress 2023
Simulation for Conference Proceedings https://doi.org/10.1109/AIIoT58121.2023.10174362
Refer here for the preprint
Warning!
The repo will not be further maintained!
Please refer to latest repo here.
Multi-camera face detection and recognition is an Artificial Intelligence (AI) based technology that leverages multiple cameras placed at different locations to detect and recognize human faces in real-world conditions accurately. While face detection and recognition technologies have exhibited high accuracy rates in controlled conditions, recognizing individuals in open environments remains challenging due to factors such as changes in illumination, movement, and occlusions. In this paper, we propose a multi-camera face detection and recognition (MCFDR) pipeline, which consists of three main parts - face detection, face recognition, and tracking. A series of model training is done with the open-source dataset to build a robust pipeline, and finally, the pipeline adopted trained YOLOv5n for face detection model with mAP 0.495, precision value of 0.868, and recall value of 0.781. The system also adopted the SphereFace SFNet20 model with an accuracy of 82.05% and a higher inference rate than SFNet64 for face recognition. These models are then fed into DeepSORT for multi-camera tracking. Our dataset has been applied to the pipeline and shows ideal outcomes with objectives achieved.
Yolov5 Object Detection with DeepSORT Tracking, using OpenSphere Face Recognition Module
- Issue 1 : Update requirements.txt
- Issue 2 : Fix video saving issue
- Issue 3 : Video stream is jumpy when track on multiple pre-recorded videos
- Issue 4 : The inference time for multi-source is wrong
- Feature 1 : Integrate screenshot face data for model retraining
Issue 2 is an existing problem in YOLOv5 repo.
However, I found that it can be avoided if you don't stop the program by keyboard interrupt (ctrl + c).
Instead, you stop the program by:
- Click on any of the windows showing the video frame
- Press 'q' (small capital letter)
- Create a conda environment, and activate it
conda create --name pipeline python=3.8.10
conda activate pipeline
- Clone the repository
git clone https://github.com/yjwong1999/Yolov5_DeepSort_Face.git
- Goto cloned folder
cd Yolov5_DeepSort_Face/Yolov5_DeepSort_Face
- Install the dependencies
pip install -r requirements.txt
- Download pretrained OpenSphere Model
sudo apt-get install zip unzip
cd opensphere/project
wget --load-cookies /tmp/cookies.txt "https://docs.google.com/uc?export=download&confirm=$(wget --quiet --save-cookies /tmp/cookies.txt --keep-session-cookies --no-check-certificate 'https://docs.google.com/uc?export=download&id=1mEAkIa9B89QzsamVhNp9mol5PZWxLthM' -O- | sed -rn 's/.*confirm=([0-9A-Za-z_]+).*/\1\n/p')&id=1mEAkIa9B89QzsamVhNp9mol5PZWxLthM" -O sfnet20_ref.zip && rm -rf /tmp/cookies.txt
unzip sfnet20_ref.zip
wget --load-cookies /tmp/cookies.txt "https://docs.google.com/uc?export=download&confirm=$(wget --quiet --save-cookies /tmp/cookies.txt --keep-session-cookies --no-check-certificate 'https://docs.google.com/uc?export=download&id=1DIrZshYMNfKCOCN54_MN1KAB0XaCWnWA' -O- | sed -rn 's/.*confirm=([0-9A-Za-z_]+).*/\1\n/p')&id=1DIrZshYMNfKCOCN54_MN1KAB0XaCWnWA" -O sfnet20_survface.zip && rm -rf /tmp/cookies.txt
unzip sfnet20_survface.zip
cd ../../
- Download test videos
wget --load-cookies /tmp/cookies.txt "https://docs.google.com/uc?export=download&confirm=$(wget --quiet --save-cookies /tmp/cookies.txt --keep-session-cookies --no-check-certificate 'https://docs.google.com/uc?export=download&id=1pDXTOr-0dYScGnifNjhHuwMz08DHl0hz' -O- | sed -rn 's/.*confirm=([0-9A-Za-z_]+).*/\1\n/p')&id=1pDXTOr-0dYScGnifNjhHuwMz08DHl0hz" -O dataset_cam1.mp4 && rm -rf /tmp/cookies.txt
wget --load-cookies /tmp/cookies.txt "https://docs.google.com/uc?export=download&confirm=$(wget --quiet --save-cookies /tmp/cookies.txt --keep-session-cookies --no-check-certificate 'https://docs.google.com/uc?export=download&id=1iND5nKvGxqGWV4EFyT8EhJKB0acqQSvF' -O- | sed -rn 's/.*confirm=([0-9A-Za-z_]+).*/\1\n/p')&id=1iND5nKvGxqGWV4EFyT8EhJKB0acqQSvF" -O dataset_cam2.mp4 && rm -rf /tmp/cookies.txt
- Find the port numbers connected with camera(s)
python3 find_port.py
- How to use
source.txt
for multli-cam or multi-streams
# In source.txt, each line should be one camera/streams
#
# 1. camera port number (which you get by running python3 find_port.py)
# 2. https or rtsp link (for online video)
# 3. <video>.mp4 or relevant video format
#
# Note that you can mix between the source format (live/video/rtsp/https)
- Do Tracking with mentioned command below
# single video cam
python3 track.py --source 0 --yolo_model yolo_face.pt --img 640 --deep_sort_model opensphere/project/sfnet20_survface --show-vid --save-vid --save-txt
# video file
python3 track.py --source dataset_cam2.mp4 --yolo_model yolo_face.pt --img 640 --deep_sort_model opensphere/project/sfnet20_survface --show-vid --save-vid --save-txt
# multi-streams (can mix between live/video/rtsp/https)
python3 track.py --source source.txt --yolo_model yolo_face.pt --img 640 --deep_sort_model opensphere/project/sfnet20_survface --show-vid --save-vid --save-txt
- Stop Tracking by
# Refer Issue 1 in TODO
#
# DO NOT: stop the program by keyboard interrupt (ctrl + c)
# DO: stop the program by:
# 1. Click on any of the windows showing the video frame
# 2. Press 'q' (small capital letter)
- Each
python3 track.py
will create an<exp index>
. You can view recordings in:
runs/track/<exp index>
- You can extract faces from recording (for retraining) by:
python3 extract_face.py --source runs/track/<exp index>
- Create a new conda environment to train OpenSphere
conda deactivate # if you are in other conda environment
conda env create -f environment.yml
conda activate opensphere
# assuming you are in Yolov5_DeepSort_Face/Yolov5_DeepSort_Face
cd opensphere
bash scripts/dataset_setup_validation_only.sh # if you only want to train with your custom dataset
OR
bash scripts/dataset_setup.sh # if you want to train with VGG Face 2 dataet
OR
bash scripts/dataset_setup_ms1m.sh # if you want to train with ms1m dataset
- Get QMUL-SurvFace dataset
cd customize
wget --load-cookies /tmp/cookies.txt "https://docs.google.com/uc?export=download&confirm=$(wget --quiet --save-cookies /tmp/cookies.txt --keep-session-cookies --no-check-certificate 'https://docs.google.com/uc?export=download&id=13ch6BPaexlKt8gXB_I8aX7p1G3yPm2Bl' -O- | sed -rn 's/.*confirm=([0-9A-Za-z_]+).*/\1\n/p')&id=13ch6BPaexlKt8gXB_I8aX7p1G3yPm2Bl" -O QMUL-SurvFace.zip && rm -rf /tmp/cookies.txt
unzip QMUL-SurvFace.zip
python3 generate_annot.py --directory QMUL-SurvFace # generate annotation for this dataset
cd ../
- Train OpenSphere Model using QMUL-SurvFace dataset
# Train SFNet20 using SphereFace loss function
CUDA_VISIBLE_DEVICES=0 python train.py --config config/train/survface_sfnet20_sphereface.yml
# Train SFNet20 using SphereFaceR loss function
CUDA_VISIBLE_DEVICES=0 python train.py --config config/train/survface_sfnet20_spherefacer.yml
# Train SFNet20 using SphereFace2 loss function
CUDA_VISIBLE_DEVICES=0 python train.py --config config/train/survface_sfnet20_sphereface2.yml
# Train SFNet64 using SphereFace loss function
CUDA_VISIBLE_DEVICES=0 python train.py --config config/train/survface_sfnet64_sphereface.yml
# Train SFNet64 using SphereFaceR loss function
CUDA_VISIBLE_DEVICES=0 python train.py --config config/train/survface_sfnet64_spherefacer.yml
# Train SFNet64 using SphereFace2 loss function
CUDA_VISIBLE_DEVICES=0 python train.py --config config/train/survface_sfnet64_sphereface2.yml
# NOTE THAT:
# CUDA_VISIBLE_DEVICES=0 means use 1st CUDA device to train
# CUDA_VISIBLE_DEVICES=0,1 means use 1st and 2nd CUDA devices to train
# and so on...
- Test OpenSphere Model using QMUL-SurvFace dataset
CUDA_VISIBLE_DEVICES=0 python test.py --config config/test/survface.yml --proj_dir project/<dir name>
- Convert OpenSphere Model to OpenVINO (for future usage)
CUDA_VISIBLE_DEVICES=0 python onnx_exporter.py --config config/test/survface.yml --proj_dir project/<dir name>
This work was supported by the Greatech Integration (M) Sdn Bhd with project number 8084-0008.
- Yolov5 + DeepSort for Object Detection and Tracking
- Yolov8 + DeepSort for Object Detection and Tracking
- OpenSphere Face Recognition
- Yolov5
- How to list available cameras OpenCV/Python
- How to wget files from Google Drive
- Technically, I used Reference [1] for detection & tracking, where Reference [2] is forked from a past version of Reference [2]
- Reference [1] and [2] assign ID based on tracks (which means each tracking has an ID, disregarding the object identity itself)
- Thus, [3] is used to learn the face identity, which is assigned as the ID for the tracking ID
@INPROCEEDINGS{10174362,
author={Wong, Yi Jie and Huang Lee, Kian and Tham, Mau-Luen and Kwan, Ban-Hoe},
booktitle={2023 IEEE World AI IoT Congress (AIIoT)},
title={Multi-Camera Face Detection and Recognition in Unconstrained Environment},
year={2023},
volume={},
number={},
pages={0548-0553},
doi={10.1109/AIIoT58121.2023.10174362}}