Modeling Temporal Concept Receptive Field Dynamically for Untrimmed Video Analysis

Related tags

Deep LearningTDCMN
Overview

Modeling Temporal Concept Receptive Field Dynamically for Untrimmed Video Analysis

This is a PyTorch implementation of the model described in our paper:

Z. Qi, S. Wang, C. Su, L. Su, W. Zhang, and Q. Huang. Modeling Temporal Concept Receptive Field Dynamically for Untrimmed Video Analysis. ACM MM 2020.

Dependencies

  • Pytorch 1.2.0
  • Cuda 9.2.148
  • Cudnn 7.6.2
  • Opencv-python 4.2.0.34
  • Python 3.6.9

Data

Dataset Prepare

  1. Download the pre-trained concept detector weights from Baidu passward 'wv0e' or Google Grive and put them in folder weights/

  2. Download the FCVID dataset from http://bigvid.fudan.edu.cn/FCVID/.

  3. The annotation information of each dataset is provided in folder data/FCVID/video_labels.

  4. Extract the video frames for each video and put the extracted frames in folder data/FCVID/frames/.

    For ActivityNet dataset ( http://activity-net.org/. ) , we use the latest released version of the dataset (v1.3).

Train

  • python main.py --gpu_ids 0,1 --model_name tdcmn_si_soa --dataset FCVID --no_test

    for other hyperparameters, please refer to opts.py file.

Test

  • Pretrained model weigths are avaiable in Baidu passward 'szlk' or Google Grive

  • Download the pre-trained weights and put them in folder results/

  • python main.py --gpu_ids 0,1 --model_name tdcmn_si_soa --dataset FCVID --resume_path pretrained_model/tdcmn_si_soa.pth --no_train --test_crop_number 1

Citation

Please cite our paper if you use this code in your own work:

@inproceedings{qi2020modeling,
  title={Modeling Temporal Concept Receptive Field Dynamically for Untrimmed Video Analysis},
  author={Qi, Zhaobo and Wang, Shuhui and Su, Chi and Su, Li and Zhang, Weigang and Huang, Qingming},
  booktitle={Proceedings of the 28th ACM International Conference on Multimedia},
  pages={3798--3806},
  year={2020}
}

Contcat

If you have any problem about our code, feel free to contact

Owner
qzhb
Video Understanding
qzhb
This is the repository of the NeurIPS 2021 paper "Curriculum Disentangled Recommendation withNoisy Multi-feedback"

Curriculum_disentangled_recommendation This is the repository of the NeurIPS 2021 paper "Curriculum Disentangled Recommendation with Noisy Multi-feedb

14 Dec 20, 2022
This project provides the proof of the uniqueness of the equilibrium and the global asymptotic stability.

Delayed-cellular-neural-network This project provides the proof of the uniqueness of the equilibrium and the global asymptotic stability. There is als

4 Apr 28, 2022
Finite-temperature variational Monte Carlo calculation of uniform electron gas using neural canonical transformation.

CoulombGas This code implements the neural canonical transformation approach to the thermodynamic properties of uniform electron gas. Building on JAX,

FermiFlow 9 Mar 03, 2022
🏎️ Accelerate training and inference of 🤗 Transformers with easy to use hardware optimization tools

Hugging Face Optimum 🤗 Optimum is an extension of 🤗 Transformers, providing a set of performance optimization tools enabling maximum efficiency to t

Hugging Face 842 Dec 30, 2022
PyTorch implementation of ShapeConv: Shape-aware Convolutional Layer for RGB-D Indoor Semantic Segmentation.

Shape-aware Convolutional Layer (ShapeConv) PyTorch implementation of ShapeConv: Shape-aware Convolutional Layer for RGB-D Indoor Semantic Segmentatio

Hanchao Leng 82 Dec 29, 2022
Code repository for "Reducing Underflow in Mixed Precision Training by Gradient Scaling" presented at IJCAI '20

Reducing Underflow in Mixed Precision Training by Gradient Scaling This project implements the gradient scaling method to improve the performance of m

Ruizhe Zhao 5 Apr 14, 2022
Implementation of Graph Convolutional Networks in TensorFlow

Graph Convolutional Networks This is a TensorFlow implementation of Graph Convolutional Networks for the task of (semi-supervised) classification of n

Thomas Kipf 6.6k Dec 30, 2022
[NeurIPS 2020] Semi-Supervision (Unlabeled Data) & Self-Supervision Improve Class-Imbalanced / Long-Tailed Learning

Rethinking the Value of Labels for Improving Class-Imbalanced Learning This repository contains the implementation code for paper: Rethinking the Valu

Yuzhe Yang 656 Dec 28, 2022
Graph Neural Networks with Keras and Tensorflow 2.

Welcome to Spektral Spektral is a Python library for graph deep learning, based on the Keras API and TensorFlow 2. The main goal of this project is to

Daniele Grattarola 2.2k Jan 08, 2023
Code for the paper "Learning-Augmented Algorithms for Online Steiner Tree"

Learning-Augmented Algorithms for Online Steiner Tree This is the code for the paper "Learning-Augmented Algorithms for Online Steiner Tree". Requirem

0 Dec 09, 2021
Baselines for TrajNet++

TrajNet++ : The Trajectory Forecasting Framework PyTorch implementation of Human Trajectory Forecasting in Crowds: A Deep Learning Perspective TrajNet

VITA lab at EPFL 183 Jan 05, 2023
Deep Networks with Recurrent Layer Aggregation

RLA-Net: Recurrent Layer Aggregation Recurrence along Depth: Deep Networks with Recurrent Layer Aggregation This is an implementation of RLA-Net (acce

Joy Fang 21 Aug 16, 2022
Implementation for On Provable Benefits of Depth in Training Graph Convolutional Networks

Implementation for On Provable Benefits of Depth in Training Graph Convolutional Networks Setup This implementation is based on PyTorch = 1.0.0. Smal

Weilin Cong 8 Oct 28, 2022
Implementation and replication of ProGen, Language Modeling for Protein Generation, in Jax

ProGen - (wip) Implementation and replication of ProGen, Language Modeling for Protein Generation, in Pytorch and Jax (the weights will be made easily

Phil Wang 71 Dec 01, 2022
Rohit Ingole 2 Mar 24, 2022
Neurons Dataset API - The official dataloader and visualization tools for Neurons Datasets.

Neurons Dataset API - The official dataloader and visualization tools for Neurons Datasets. Introduction We propose our dataloader API for loading and

1 Nov 19, 2021
Machine learning framework for both deep learning and traditional algorithms

NeoML is an end-to-end machine learning framework that allows you to build, train, and deploy ML models. This framework is used by ABBYY engineers for

NeoML 704 Dec 27, 2022
Tutorial page of the Climate Hack, the greatest hackathon ever

Tutorial page of the Climate Hack, the greatest hackathon ever

UCL Artificial Intelligence Society 12 Jul 02, 2022
Establishing Strong Baselines for TripClick Health Retrieval; ECIR 2022

TripClick Baselines with Improved Training Data Welcome 🙌 to the hub-repo of our paper: Establishing Strong Baselines for TripClick Health Retrieval

Sebastian Hofstätter 3 Nov 03, 2022
Code for the ICASSP-2021 paper: Continuous Speech Separation with Conformer.

Continuous Speech Separation with Conformer Introduction We examine the use of the Conformer architecture for continuous speech separation. Conformer

Sanyuan Chen (陈三元) 81 Nov 28, 2022