Hybrid multi scale hard switch YOLOv4 network for cricket video summarization

نویسندگان

چکیده

Cricket is a popular sport with lengthy duration that makes it challenging to watch in its entirety. Therefore, video summarization techniques are essential providing viewers condensed version of the match's exciting moments. Automated cricket difficult due sport's regulations and extended sessions. Existing methods often include repetitive shots, making summary less concise informative. this paper proposes hybrid framework uses audio text features extract clips from raw video. The employs Multi-Scale Hard Switch YOLOv4 (MSHS-YOLOv4) network accurately detect label events, including small details such as ball hitting stumps. A significance score computed for each event generate includes most significant proposed method eliminates replay reducing redundancy more concise. combines identify moments, MSHS-YOLOv4 computes event, shots summary. outperforms existing terms accuracy, precision, recall, F1-score, error. analysis shows increase performance compared methods.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Graph Hybrid Summarization

One solution to process and analysis of massive graphs is summarization. Generating a high quality summary is the main challenge of graph summarization. In the aims of generating a summary with a better quality for a given attributed graph, both structural and attribute similarities must be considered. There are two measures named density and entropy to evaluate the quality of structural and at...

متن کامل

A Unified Multi-Faceted Video Summarization System

T his paper addresses automatic summarization and search in visual data comprising of videos, live streams and image collections in a unified manner. In particular, we propose a framework for multi-faceted summarization which extracts keyframes (image summaries), skims (video summaries) and entity summaries (summarization at the level of entities like objects, scenes, humans and faces in the vi...

متن کامل

Multi-Scale Video Frame-Synthesis Network with Transitive Consistency Loss

Traditional approaches to interpolate/extrapolate frames in a video sequence require accurate pixel correspondences between images, e.g., using optical flow. Their results stem on the accuracy of optical flow estimation, and could generate heavy artifacts when flow estimation failed. Recently methods using auto-encoder has shown impressive progress, however they are usually trained for specific...

متن کامل

A Multi-scale Multiple Instance Video Description Network

Generating natural language descriptions for in-thewild videos is a challenging task. Most state-of-the-art methods for solving this problem borrow existing deep convolutional neural network (CNN) architectures (Alexnet, Googlenet) to extract a visual representation of the input video. However, these deep CNN architectures are designed for single-label centered-positioned object classification....

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Wireless Networks

سال: 2023

ISSN: ['1572-8196', '1022-0038']

DOI: https://doi.org/10.1007/s11276-023-03449-8