This project uses this fork of Facebook's MMF framework.
!!! The DATASET is located in this repository under the following paths:
/data/datasets/memes_tgt/defaults/annotations (target labels) dataset folder is same as the other task
!!! We are using the open source Multimodal Framework developed by FacebookResearch: https://mmf.sh/
- Prerequisites - generating image caption features for VisualBERT and ViLBERT:
- Clone this repository and go to './mmf' path
- Install MMF according to the instructions here: https://mmf.readthedocs.io/en/website/notes/installation.html
- Install the following packages: 'pip install yacs, opencv-python, cython' (if using 'pip', any package manager works)
- Clone vqa-maskrcnn-benchmark repository: https://gitlab.com/vedanuj/vqa-maskrcnn-benchmark
- Run 'python setup.py build'
- Run 'python setup.py develop'
- Run the feature extraction script from the following path: 'mmf/tools/scripts/features/extract_features_vmb.py'
- After feature extraction is done convert the features to a .mdb file with the following script: 'mmf/tools/scripts/features/extract_features_vmb.py'
- Rename the .mdb features file to 'deceptron.lmdb' and move it to '/root/.cache/torch/mmf/data/datasets/memes/defaults/features/'
- Running the models - open 'HarMeme.ipynb' and run the code inside.