forked from Chinmay26/Near-Duplicate-Video-Detection
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Showing
1 changed file
with
29 additions
and
2 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,2 +1,29 @@ | ||
# Near-Duplicate-Video-Detection | ||
Detecting near-duplicate videos by aggregating features from intermediate CNN layers. This is a WIP | ||
# Near-Duplicate-Video-Detection (NDVD) | ||
With the explosion of social networks, video content has risen exponentially over the last few years. Currently, Youtube reports 500h of video content being uploaded every minute. A Cisco forecast estimates that videos will constitute 80% of internet traffic by 2019. According to a recent study, 31.7% of videos on Youtube are duplicates with duplicates occupying 24% of storage space. It has become essential that we build robust systems to detect and purge these duplicates. | ||
|
||
Near duplicate videos are a bigger class of problems which cover duplicate videos. Near duplicates videos are videos which have same visual content but differ in format ( scale, transformation, encoding etc ) or have small content modifications ( color / lightning / small text superimposition etc). | ||
|
||
### Problem Statement | ||
Given a corpus of videos, identify if a query video is a near duplicate of a video in corpus. | ||
|
||
## Dataset | ||
The standard for NDVD tasks is the [CC_WEB_VIDEO](http://vireo.cs.cityu.edu.hk/webvideo/) dataset. The dataset contains a total of 13,129 videos and 24 query videos. The dataset already provides 398,008 keyframes extracted from the videos by shot boundary detection method. | ||
|
||
## Approach | ||
The problem is tackled by using a bag-of visual words model for each video. This is done in following phases | ||
- Feature Extraction: Convolution Neural Networks are used to extract features from the video keyframes. Pretrained CNN networks have been proven to work well on many vision tasks such as Classification, Segmentation etc. Here, pretrained weights of AlexNet is used to extract feature vectors. Each video keyframe is forward passed through the intermediate layers of AlexNet to get frame feature vector. Max-pooling is applied on the intermediate feature maps to extract one single value. Each frame is then represented by a feature vector of 1376 dimensions.The video-level feature vector is calculated by summing over the individual video keyframes. | ||
|
||
|
||
- Visual Codebook Generation: A visual codebook is generated using the above feature vectors. Online Mini-batch K means is used to generate the codebook clusters. A sample of random 100K frames are used for visual codebook generation. K = 1000 gives best results for the NDVD task. | ||
|
||
- Video level histogram: For each keyframe in a video, the Nearest cluster is identified to generate a keyframe level histogram. Correspondingly, a video level histogram is generated by summing oveer the individual keyframe histograms. | ||
|
||
- | ||
|
||
## Challenges | ||
- Scalability: How | ||
|
||
## References: | ||
[Near-duplicate Video Detection using CNN intermediate features](https://link.springer.com/chapter/10.1007/978-3-319-51811-4_21) | ||
[NDVD using metric learning](openaccess.thecvf.com/content_ICCV_2017_workshops/papers/w5/Kordopatis-Zilos_Near-Duplicate_Video_Retrieval_ICCV_2017_paper.pdf) | ||
|