Skip to content

OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark

License

Notifications You must be signed in to change notification settings

owaisCS/VARB-mmaction2

 
 

Repository files navigation

Video Action Recognition Benchmark (VARB)

Video Action Recognition & Need for a Benchmark

Video action recognition is one of the major challenges in visual scene understanding. With the advent of deep learning (DL), research in this field has witnessed significant improvements over the last decade where the deep models learn higher dimensional video representation that helps classifying actions. Although there has been a remarkable progress, the lack of any unified evaluation makes it difficult to gauge the potential of the various contributing factors viz. sensitive hyperparameters, pretraining datasets, size of downstream datasets and transfer learning algorithm (full-network, linear or limited fine-tuning).

What is VARB?

  • VARB stands for Video Action Recognition Benchmark
  • VARB provides a comprehensive study of transfer learning for video action recognition
  • Through VARB, we analyze $6$ different video activity recognition models, starting from the simpler ones like $2$-D CNNs to the more intricate ones like $3$-D CNNs and transformer architectures ($2$-D and $3$-D).
  • For evaluation, VARB uses $14$ publicly available action recognition datasets with videos from $7$ distinct domains namely daily actions, sports, actions in the dark, indoor actions, masked-depth actions and construction actions.
  • VARB introduces a new dataset Construction-Actions by trimming and manually annotating activity videos of construction sites from YouTube-8M.
  • VARB bridges a significant gap in video action recognition research by bringing different action recognition models and datasets to a standard benchmark

Requirements

Setup

conda create --name open-mmlab python==3.8
conda activate avatar_env
pip install -r requirements.txt

conda activate open-mmlab

Results

The performance of the

About

OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 97.9%
  • Shell 1.7%
  • Other 0.4%