Indexed Dataset from YouTube for a Content-Based Video Search Engine

Adly, Ahmad and Hegazy, Islam and Elarif, Taha and Abdelwahab, M. S. (2021) Indexed Dataset from YouTube for a Content-Based Video Search Engine. International Journal of Intelligent Computing and Information Sciences, 21 (1). pp. 196-215. ISSN 2535-1710

[thumbnail of IJICIS_Volume 21_Issue 1_Pages 196-215.pdf] Text
IJICIS_Volume 21_Issue 1_Pages 196-215.pdf - Published Version

Download (1MB)

Abstract

Numerous researches on content-based video indexing and retrieval besides video search engines are tied to a large-scaled video dataset. Unfortunately, reduction in open-sourced datasets resulted in complications for novel approaches exploration. Although, video datasets that index video files located on public video streaming services have other purposes, such as annotation, learning, classification, and other computer vision areas, with little interest in indexing public video links for purpose of searching and retrieval. This paper introduces a novel large-scaled dataset based on YouTube video links to evaluate the proposed content-based video search engine, gathered 1088 videos, that represent more than 65 hours of video, 11,000 video shots, and 66,000 unmarked and marked keyframes, 80 different object names used for marking. Moreover, a state-of-the-art features vector, and combinational-based matching, beneficial to the accuracy, speed, and precision of the video retrieval process. Any video record in the dataset is represented by three features: temporal combination vector, object combination vector with shot annotations, and 6 keyframes, sideways with other metadata. Video classification for the dataset was also imposed to expand the efficiency of retrieval of video-based queries. A two-phased approach has been used based on object and event classification, storing video records in aggregations related to feature vectors extracted. While object aggregation stores video records with the maximal occurrence of extracted object/concept from all shots, event aggregation classify based on groups according to the number of shots per video. This study indexed 58 out of 80 different object/concept categories, each has 9 shot number groups.

Item Type: Article
Subjects: STM Digital > Computer Science
Depositing User: Unnamed user with email support@stmdigital.org
Date Deposited: 27 Jun 2023 06:54
Last Modified: 21 Sep 2024 04:50
URI: http://research.asianarticleeprint.com/id/eprint/1234

Actions (login required)

View Item
View Item