# ECO-efficient-video-understanding **Repository Path**: Chris_Ch0u/ECO-efficient-video-understanding ## Basic Information - **Project Name**: ECO-efficient-video-understanding - **Description**: No description available - **Primary Language**: Unknown - **License**: MIT - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2018-08-04 - **Last Updated**: 2020-12-18 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README #### Code and models of [paper](https://arxiv.org/pdf/1804.09066.pdf). " ECO: Efficient Convolutional Network for Online Video Understanding" By Mohammadreza Zolfaghari, Kamaljeet Singh, Thomas Brox ### Update - **2018.8.01**: Scripts for online recognition and video captioning - **2018.7.30**: Adding codes and models - **2018.4.17**: Repository for ECO. ### Introduction This repository will contains all the required models and scripts for the paper [ECO: Efficient Convolutional Network for Online Video Understanding](https://arxiv.org/pdf/1804.09066.pdf). ![](doc_files/s_model.png) In this work, we introduce a network architecture that takes long-term content into account and enables fast per-video processing at the same time. The architecture is based on merging long-term content already in the network rather than in a post-hoc fusion. Together with a sampling strategy, which exploits that neighboring frames are largely redundant, this yields high-quality action classification and video captioning at up to 230 videos per second, where each video can consist of a few hundred frames. The approach achieves competitive performance across all datasets while being 10x to 80x faster than state-of-the-art methods. ### Results Action Recognition on UCF101 and HMDB51 | Video Captioning on MSVD dataset :-------------------------:|:-------------------------: ![](doc_files/s_fig1.png) | ![](doc_files/s_fig2.png) ### Online Video Understanding Results Model trained on UCF101 dataset | Model trained on Something-Something dataset :-------------------------:|:-------------------------: ![](doc_files/uc_gif1.gif) | ![](doc_files/sm_gif1.gif) ### Requirements 1. Requirements for `Python` 2. Requirements for `Caffe` (see: [Caffe installation instructions](http://caffe.berkeleyvision.org/installation.html)) ### Installation Build Caffe ```Shell cd $caffe_FAST_ROOT/ # Now follow the Caffe installation instructions here: # http://caffe.berkeleyvision.org/installation.html make all -j8 ``` ### Usage *After successfully completing the [installation](#installation)*, you are ready to run all the following experiments. ### Training 1. Download the initialization and trained models: ```Shell sh download_models.sh ``` 2. Train ECO Lite on kinetics dataset: sh models_ECO_Lite/kinetics/run.sh ### TODO 1. Data 2. Tables and Results 3. Demo 4. PyTorch version of ECO ### Contact [Mohammadreza Zolfaghari](https://github.com/mzolfaghari/ECO_efficient_video_understanding) Questions can also be left as issues in the repository. We will be happy to answer them.