Abstract
Automatic video summarization aims to provide brief representa-
tion of videos. Its evaluation is quite challenging, usually relying on comparison
with user summaries. This study views it in a diferent perspective in terms
of verifying the consistency of user summaries, as the outcome of video sum-
marization is usually judged based on them. We focus on human consistency
evaluation of static video summaries in which the user summaries are evalu-
ated among themselves using the consistency modelling method we proposed
recently. The purpose of such consistency evaluation is to check whether the
users agree among themselves. The evaluation is performed on diferent pub-
licly available datasets. Another contribution lies in the creation of static video
summaries from the available video skims of the SumMe datatset. The results
show that the level of agreement varies signigcantly between the users for the
selection of key frames, which denotes the hidden challenge in automatic video
summary evaluation. Moreover, the maximum agreement level of the users for
a certain dataset, may indicate the best performance that the automatic video
summarization techniques can achieve using that dataset.
Original language | English |
---|---|
Pages (from-to) | 12281-12306 |
Number of pages | 26 |
Journal | Multimedia Tools and Applications |
Volume | 78 |
Early online date | 20 Oct 2018 |
DOIs | |
Publication status | E-pub ahead of print - 20 Oct 2018 |
Keywords
- Video Summarization
- Keyframe Extraction
- PerformanceEvaluation
- Consistency modelling
- User Consistency
- Performance evaluation
- User consistency
- Video summarization
- Keyframe extraction