AI4Good project for detecting waste in environment. www.detectwaste.ml.
Our latest results were published in Waste Management journal in article titled Deep learning-based waste detection in natural and urban environments.
You can find more technical details in our technical report Waste detection in Pomerania: non-profit project for detecting waste in environment.
Did you know that we produce 300 million tons of plastic every year? And only the part of it is properly recycled.
The idea of detect waste project is to use Artificial Intelligence to detect plastic waste in the environment. Our solution is applicable for video and photography. Our goal is to use AI for Good.
In Detect Waste in Pomerania project we used 9 publicity available datasets, and additional data collected using Google Images Download.
For more details, about the data we used, check our jupyter notebooks with data exploratory analysis.
Data annotations: https://github.com/wimlds-trojmiasto/detect-waste/tree/main/annotations
-
TACO bboxes - in progress. TACO dataset can be downloaded here. TACO bboxes will be avaiable for download soon.
Clone Taco repository
git clone https://github.com/pedropro/TACO.git
Install requirements
pip3 install -r requirements.txt
Download annotated data
python3 download.py
-
Clone UAVVaste repository
git clone https://github.com/UAVVaste/UAVVaste.git
Install requirements
pip3 install -r requirements.txt
Download annotated data
python3 main.py
-
Download directly from web
wget https://conservancy.umn.edu/bitstream/handle/11299/214865/dataset.zip?sequence=12&isAllowed=y
-
Download directly from web
wget https://conservancy.umn.edu/bitstream/handle/11299/214366/trash_ICRA19.zip?sequence=12&isAllowed=y
-
Download directly from google drive
-
In order to download you must first authenticate using a kaggle API token. Read about it here
kaggle datasets download -d arkadiyhacks/drinking-waste-classification
-
Clone wade-ai repository
git clone https://github.com/letsdoitworld/wade-ai.git
For coco annotation check: majsylw/wade-ai/tree/coco-annotation
-
TrashNet - The dataset spans six classes: glass, paper, cardboard, plastic, metal, and trash.
Clone trashnet repository
git clone https://github.com/garythung/trashnet
-
waste_pictures - The dataset contains ~24k images grupped by 34 classes of waste for classification purposes.
In order to download you must first authenticate using a kaggle API token. Read about it here
kaggle datasets download -d wangziang/waste-pictures
For more datasets check: waste-datasets-review
To train only on TACO dataset with detect-waste classes:
-
run annotations_preprocessing.py
python3 annotations_preprocessing.py
new annotations will be saved in annotations/annotations_train.json and annotations/annotations_test.json
For binary detection (litter and background) check also generated new annotations saved in annotations/annotations_binary_train.json and annotations/annotations_binary_test.json.
To train on one or multiple datasets on a single class:
-
run annotations_preprocessing_multi.py
python3 annotations_preprocessing_multi.py
new annotations will be split and saved in annotations/binary_mixed_train.json and annotations/binary_mixed_test.json
Example bash file is in annotations_preprocessing_multi.sh and can be run by
bash annotations_preprocessing_multi.sh
Script will automatically split all datasets to train and test set with MultilabelStratifiedShuffleSplit
. Then it will convert datasets to one class - litter. Finally all datasets will be concatenated to form single train and test files annotations/binary_mixed_train.json
and annotations/binary_mixed_test
.
For more details check annotations directory.
To read more about past waste detection works check litter-detection-review.
-
To train EfficientDet check
efficientdet/README.md
To train EfficientDet implemented in Pytorch Lightning check branch
effdet_lightning
We based our implementation on efficientdet-pytorch by Ross Wightman.
-
To train detr check
detr/README.md
(WIP)PyTorch training code and pretrained models for DETR (DEtection TRansformer). Authors replaced the full complex hand-crafted object detection pipeline with a Transformer, and matched Faster R-CNN with a ResNet-50, obtaining 42 AP on COCO using half the computation power (FLOPs) and the same number of parameters. Inference in 50 lines of PyTorch.
For implementation details see End-to-End Object Detection with Transformers by Facebook.
-
To train Mask R-CNN check
MaskRCNN/README.md
Our implementation based on tutorial.
-
To train Faster R-CNN on TACO dataset check
FastRCNN/README.md
-
To train choosen model check
classifier/README.md
- Waste detection using EfficientDet
In our github repository you will find EfficientDet code already adjusted for our mixed dataset. To run training for single class just clone repository, move to efficientdet directory, install necessary dependencies, and launch train.py
script with adjusted parameters, like: path to images, path to directory with annotations (you can use ours provided in annotations directory), model parameters and its specific name. It can be done as in the example below.
python3 train.py path_to_all_images \
--ann_name ../annotations/binary_mixed --model tf_efficientdet_d2 \
--batch-size 4 --decay-rate 0.95 --lr .001 --workers 4 --warmup-epochs 5 \
--model-ema --dataset multi --pretrained --num-classes 1 --color-jitter 0.1 \
--reprob 0.2 --epochs 20 --device cuda:0
- Waste classification using EfficientNet
In this step switch to classifier directory. At first just crop waste objects from images of waste (the same as in previous step).
python3 cut_bbox_litter.py --src_img path_to_whole_images \
--dst_img path_to_destination_directory_for_images \
--square --zoom 1
In case of using unlabelled OpenLitterMap dataset, make pseudo-predictions using previously trained EfficientDet and map them with orginal openlittermap annotations.
python3 sort_openlittermap.py \
--src_ann path_to_original_openlittermap_annotations \
--coco path_to_our_openlittermap_annotations \
--src_img path_to_whole_images \
--dst_img path_to_destination_directory_for_images
To run classifier training in command line just type:
python train_effnet.py --data_img path/to/images/train/ \
--save path/to/checkpoint.ckpt \
--model efficientnet-b2 \
--gpu 0 \
--pseudolabel_mode per-batch
We provided make_predictions.py
script to draw bounding boxes on choosen image. For example script can be run on GPU (id=0) with arguments:
python make_predictions.py --save directory/to/save/image.png \
--detector path/to/detector/checkpoint.pth \
--classifier path/to/clasifier/checkpoint.pth \
--img path/or/url/to/image --device cuda:0
or on video with --video
argument:
python make_predictions.py --save directory/to/save/frames \
--detector path/to/detector/checkpoint.pth \
--classifier path/to/clasifier/checkpoint.pth \
--img path/to/video.mp4 --device cuda:0 --video \
--classes label0 label1 label2
If you managed to process all the frames, just run the following command from the directory where you saved the results:
ffmpeg -i img%08d.jpg movie.mp4
For experiment tracking we mostly used neptune.ai. To use Neptune
follow the official Neptune tutorial on their website:
-
Log in to your account
-
Find and set Neptune API token on your system as environment variable (your NEPTUNE_API_TOKEN should be added to ~./bashrc)
-
Add your project_qualified_name name in the
train_<net_name>.py
neptune.init(project_qualified_name = 'YOUR_PROJECT_NAME/detect-waste')
Currently it is set to a private detect-waste neptune space.
-
install neptune-client library
pip install neptune-client
For more check LINK.
model | backbone | Dataset | # classes | bbox AP@0.5 | bbox AP@0.5:0.95 | mask AP@0.5 | mask AP@0.5:0.95 |
---|---|---|---|---|---|---|---|
DETR | ResNet 50 | TACO bboxes | 1 | 46.50 | 24.35 | x | x |
DETR | ResNet 50 | TACO bboxes | 7 | 12.03 | 6.69 | x | x |
DETR | ResNet 50 | * Multi |
1 | 50.68 | 27.69 | ** 54.80 |
** 32.17 |
DETR | ResNet 101 | * Multi |
1 | 51.63 | 29.65 | 37.02 | 19.33 |
Mask R-CNN | ResNet 50 | * Multi |
1 | 27.95 | 16.49 | 23.05 | 12.94 |
Mask R-CNN | ResNetXt 101 | * Multi |
1 | 19.70 | 6.20 | 24.70 | 13.20 |
EfficientDet-D2 | EfficientNet-B2 | Taco bboxes | 1 | 61.05 | x | x | x |
EfficientDet-D2 | EfficientNet-B2 | Taco bboxes | 7 | 18.78 | x | x | x |
EfficientDet-D2 | EfficientNet-B2 | Drink-waste | 4 | 99.60 | x | x | x |
EfficientDet-D2 | EfficientNet-B2 | MJU-Waste | 1 | 97.74 | x | x | x |
EfficientDet-D2 | EfficientNet-B2 | TrashCan v1 | 8 | 91.28 | x | x | x |
EfficientDet-D2 | EfficientNet-B2 | Wade-AI | 1 | 33.03 | x | x | x |
EfficientDet-D2 | EfficientNet-B2 | UAVVaste | 1 | 79.90 | x | x | x |
EfficientDet-D2 | EfficientNet-B2 | Trash ICRA19 | 7 | 9.47 | x | x | x |
EfficientDet-D2 | EfficientNet-B2 | * Multi |
1 | 74.81 | x | x | x |
EfficientDet-D3 | EfficientNet-B3 | * Multi |
1 | 74.53 | x | x | x |
*
Multi - name for mixed open dataset (with listed below datasets) for detection/segmentation task**
results achived with frozeen weights from detection task (after addition of mask head)
Pretrained weights of the best efficientdet model are available to download here: https://drive.google.com/drive/u/0/folders/1wNWoH8rdkG05sBw-OCXp3J73uJPxhcxH
model | # classes | ACC | sampler | pseudolabeling |
---|---|---|---|---|
EfficientNet-B2 | 8 | 73.02 | Weighted | per batch |
EfficientNet-B2 | 8 | 74.61 | Random | per epoch |
EfficientNet-B2 | 8 | 72.84 | Weighted | per epoch |
EfficientNet-B4 | 7 | 71.02 | Random | per epoch |
EfficientNet-B4 | 7 | 67.62 | Weighted | per epoch |
EfficientNet-B2 | 7 | 72.66 | Random | per epoch |
EfficientNet-B2 | 7 | 68.31 | Weighted | per epoch |
EfficientNet-B2 | 7 | 74.43 | Random | None |
ResNet-50 | 8 | 60.60 | Weighted | None |
- 8 classes - 8th class for additional background category
- we provided 2 methods to update pseudo-labels: per batch and per epoch
@article{MAJCHROWSKA2022274,
title = {Deep learning-based waste detection in natural and urban environments},
journal = {Waste Management},
volume = {138},
pages = {274-284},
year = {2022},
issn = {0956-053X},
doi = {https://doi.org/10.1016/j.wasman.2021.12.001},
url = {https://www.sciencedirect.com/science/article/pii/S0956053X21006474},
author = {Sylwia Majchrowska and Agnieszka Mikołajczyk and Maria Ferlin and Zuzanna Klawikowska
and Marta A. Plantykow and Arkadiusz Kwasigroch and Karol Majek},
keywords = {Object detection, Semi-supervised learning, Waste classification benchmarks,
Waste detection benchmarks, Waste localization, Waste recognition},
}
@misc{majchrowska2021waste,
title={Waste detection in Pomerania: non-profit project for detecting waste in environment},
author={Sylwia Majchrowska and Agnieszka Mikołajczyk and Maria Ferlin and Zuzanna Klawikowska
and Marta A. Plantykow and Arkadiusz Kwasigroch and Karol Majek},
year={2021},
eprint={2105.06808},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
├── LICENSE
├── README.md
| <- The top-level README for developers using this project.
├── annotations <- annotations in json
│
├── classifier <- implementation of CNN for litter classification
│
├── detr <- implementation of DETR for litter detection
│
├── efficientdet <- implementation of EfficientDet for litter detection
│
├── fastrcnn <- implementation of FastRCNN for litter segmentation
│
├── maskrcnn <- implementation of MaskRCNN for litter segmentation
│
├── notebooks <- jupyter notebooks.
│
├── utils <- source code with useful functions
│
├── requirements.txt <- The requirements file for reproducing the analysis environment, e.g.
│ generated with `pip freeze > requirements.txt`
├── setup.py <- makes project pip installable (pip install -e .) so src can be imported
├── src <- Source code for use in this project.