Modeling media memorability has been a consistent challenge in the field of machine learning. The Predicting Media Memorability task in MediaEval2020 is the latest benchmark among similar challenges addressing this topic. Building upon techniques developed in previous iterations of the challenge, we developed ensemble methods with the use of extracted video, image, text, and audio features. Critically, in this work we introduce and demonstrate the efficacy and high generalizability of extracted audio embeddings as a feature for the task of predicting media memorability.
Download training set from google drive and unzip in root
Create conda environnment with conda env create -f environment.yml
Activate conda environment with conda activate video-mem
Run setup with python setup.py
(downloads videos and extracts audio)
Audio extraction requires command line access to FFmpeg
conda activate video-mem && jupyter notebook .
- training_set/ #Data belongs here
- Features/
- Videos/ #Created by setup.py
- Audio/ #Created by setup.py
- video_urls.csv
- ...
- setup.py #Downloads videos
- README.md
- environment.yml
- train.ipynb #Training workflow
Team | Spearman | Pearson | MSE |
---|---|---|---|
Memento10k | 0.137 | 0.13 | 0.01 |
UC Berkeley (Us!) | 0.136 | 0.145 | 0.01 |
MeMAD | 0.101 | 0.09 | 0.01 |
KT-UPB | 0.053 | 0.085 | 0.01 |
Essex-NLIP | 0.042 | 0.042 | 0.01 |
DCU@ML-Labs | 0.034 | 0.078 | 0.10 |
Average | 0.058 | 0.066 | 0.013 |
Variance | 0.002 | 0.002 | 0.000 |