Back close

Visual recognition and classification of videos using deep convolutional neural networks

Publication Type : Journal Article

Publisher : International Journal of Engineering and Technology(UAE),

Source : International Journal of Engineering and Technology(UAE), Volume 7, Issue 2, p.85-88 (2018)

Url : https://www.scopus.com/record/display.uri?eid=2-s2.0-85047852784&origin=resultslist

Keywords : Bag of words features, Convolutional neural networks, K-Means clustering, Local binary patterns, Sports videos, SURF, Video processing

Campus : Mysuru

School : School of Arts and Sciences

Department : Computer Science

Year : 2018

Abstract : Classification of videos based on its content is one of the challenging and significant research problems. In this paper, a simple and efficient model is proposed for classification of sports videos using deep learned convolution neural networks. In the proposed research, the gray scale variants of image frames are employed for classification process through convolution technique at varied levels of abstraction by adapting it through a sequence of hidden layers. The image frames considered for classification are obtained after the duplicate frame elimination and each frame is further rescaled to dimension 120×240. The sports videos categories used for experimentation include badminton, football, cricket and tennis which are downloaded from various sources of google and YouTube. The classification in the proposed method is performed with Deep Convolution Neural Networks (DCNN) with around 20 filters each of size 5×5 with around stride length of2 and its outcomes are compared with Local Binary Patterns (LBP), Bag of Words Features (BWF) technique. The SURF features are extracted from the BWF technique and further 80% of strongest feature points are employed for clustering the image frames using K-Means clustering technique with an average accuracy achieved of about 87% in classification. The LBF technique had produced an average accuracy of 73% in differentiating one image frame to other whereas the DCNN had shown a promising outcome with accuracy of about 91% in case of 40% training and 60% test datasets, 99% accuracy in case of 60% training an 40% test datasets. The results depict that the proposed method outperforms the image processing-based techniques LBP and BWF. © 2018 Authors.

Cite this Research Publication : Shobha Rani N., Pramod Rao, and Paul Clinton, “Visual recognition and classification of videos using deep convolutional neural networks”, International Journal of Engineering and Technology(UAE), vol. 7, no. 2, pp. 85-88, 2018.

Admissions Apply Now