International Research Journal of Engineering and Technology (IRJET)
e-ISSN: 2395-0056
Volume: 07 Issue: 12 | Dec 2020
p-ISSN: 2395-0072
www.irjet.net
Real Time Video to Text Summarization using Neural Network Abhishek Yadav1, Anjali Vishwakarma2, Shyama Panickar3, Prof. Satish Kuchiwale4 1-3Students,
Dept. of Computer Science, Smt. Indira Gandhi College of Engineering, Navi Mumbai, Maharashtra, India 4Teacher, Dept. of Computer Science, Smt. Indira Gandhi College of Engineering, Navi Mumbai, Maharashtra, India ---------------------------------------------------------------------***----------------------------------------------------------------------
Abstract - This paper represents a model for automatically identifying the important parts of a real time video and annotating the video with captions to enable a rich and more concise condensation of the video. The main idea is to select the key frames from the live video feed, and caption them for text summarization. On the other hand, captioned image can help a video captioning model to learn better semantic representations. This model proposes a general neural network configuration that jointly considers two supervisory signals in the training phase and generates both video summary and corresponding captions for a given images in the test phase. Key Words: Caption signals, video captioning model, Semantic representations. 1. INTRODUCTION The prevalence of recording devices encourages more people to capture their daily life with video data content. But, the large amount of video data makes it more difficult to navigate, particularly long videos such as surveillance videos or CCTV footages. For larger videos, automatically identifying the important parts/frames of the video content and enabling them with captions will give a richer and more concise condensation of the video. It is still time-consuming for users to navigate or to search through a summarized video. So, the automatic video summarization has been proposed to extract a compact representation of the video data into textual form. The proposed system offers a brief semantic understanding of a long video just through a text summary. 2. RELATED WORK 2.1 Video Description Traditionally, researchers use unsupervised methods for automatic real-time video summarization. This is done either by taking a holistic view of the entire video or by identifying the local differentiation among the adjacent frames. In these methods, researchers design criteria such as relevance, diversity, and representativeness to select important frames or shots from the video. Some researchers utilize web media and metadata as prior knowledge to generate better summarization results. Visual attention is also used to select important frames. However, video summarization requires a semantical understanding of the video content and is hard to model with a heuristic design. Our system also uses a supervised method with RNN for video summarization I.e. LSTM. Different from previous work, we incorporate the context of a video captioning model to learn a more semantic- driven video representation for the summarization task. 2.2 Image Captioning Video understanding is an important subject in the field of computer vision. Researchers have worked on many different topics of video content analysis such as video classification, action recognition, and video captioning. Recently, with the success of image captioning using CNN and RNN models, many researchers have adopted similar approaches to describe video content with natural language. There are methods that emphasize caption generation including: Semantic supervision which designs auxiliary objectives to exploit visual semantic concepts to improve captioning quality. (2) Approaches to mitigate the objective mismatch problem. 3. METHODOLOGY In this section, we have discussed the methodology for real-time video to text summarization using deep neural network-based models. We have introduced a model for feature extraction and image captioning. The initial phase is the selection of the keyframes from the real-time video feed at every interval. For the image feature extraction, we have used the Convolutional Neural Network (CNN). To train the model on the Flickr 8k dataset first we have to extract the features from the images by using the inceptionxv3 model and then we have to extract the captions for the images also. The captions will need to be converted to numbers before presenting to the model. The first step in encoding the captions is to create a consistent mapping
Š 2020, IRJET
|
Impact Factor value: 7.529
|
ISO 9001:2008 Certified Journal
|
Page 1828