Skip to content

xiehaopeng/IIITC

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 

Repository files navigation

IIITCNET: Intra- and Inter-frame Iterative Temporal Convolutional Networks for Video Stabilization

Official project page for paper, IIITCNET: Intra- and Inter-frame Iterative Temporal Convolutional Networks for Video Stabilization.

we show an example videos to compare our work with other methods. result demo

The video can be obtained here .

Dataset

dataset demo

You can download our video stabilization dataset from Google Drive

We provide a real-world video dataset of the pre-print paper 'IIITCNET: Intra- and Inter-frame Iterative Temporal Convolutional Networks for Video Stabilization'.

There are 140 pairs of synchronized videos in this dataset. Each pair of videos includes a jitter video and a stabilized video after hardware stabilization. The length of each video is about 15-30 seconds and frames per second is 30. The resolution is 1920x1080.

According to the motion type, jitter degree and challenging content, the dataset was divided into the following 9 categories:

  1. simple. It only contains the linear motion of the lens, and the scene content is simple.
  2. running. It has the violent up and down motion while shaking.
  3. quick rotation. It contains blurring and distortion caused by the quick rotation of the lens
  4. vehicle. The jitter from vehicle motion was added in videos.
  5. parallax. It has big difference in the apparent position of an object viewed along two different lines of sight.
  6. depth. The depth transformation in the video is discontinuous.
  7. occlusion. It includes large-scale occlusion caused by close-up objects or obstacles.
  8. crowd. It has a large number of moving objects in different motion states.
  9. low-quality. This kind of video will cause feature extraction failures. It was divided into 4 sub-categories:dark, blur, noise and watermark.

Usage of this repo

Prerequisites

  • Python 3.5
  • CUDA 9.0
  • pytorch 1.1.0
  • torchvision 0.3.0
  • cudatoolkit 9.0
  • numpy
  • ...

Data preparation

You can download deepstab (7.9GB) or our dataset (10.2GB) as training data.

We use the public bundled camera path dataset in our experiment. You can also use your own shake videos to verify the effect of the model.

After downloading, You need to modify the video path of the code in every ".py" file.

Training

You need to modify the save path of the model, which is in the “./model” folder by default.

python train_model.py

Testing

Before testing, you can train the model yourself or use the model parameters we have trained. our dataset

python test_model.py

Contact

Please contact the first author Haopeng Xie ([email protected]) for any questions.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages