• English
    • Ελληνικά
    • Deutsch
    • français
    • italiano
    • español
  • français 
    • English
    • Ελληνικά
    • Deutsch
    • français
    • italiano
    • español
  • Ouvrir une session
Voir le document 
  •   Accueil de DSpace
  • Επιστημονικές Δημοσιεύσεις Μελών ΠΘ (ΕΔΠΘ)
  • Δημοσιεύσεις σε περιοδικά, συνέδρια, κεφάλαια βιβλίων κλπ.
  • Voir le document
  •   Accueil de DSpace
  • Επιστημονικές Δημοσιεύσεις Μελών ΠΘ (ΕΔΠΘ)
  • Δημοσιεύσεις σε περιοδικά, συνέδρια, κεφάλαια βιβλίων κλπ.
  • Voir le document
JavaScript is disabled for your browser. Some features of this site may not work without it.
Tout DSpace
  • Communautés & Collections
  • Par date de publication
  • Auteurs
  • Titres
  • Sujets

Summarization of User-Generated Videos Fusing Handcrafted and Deep Audiovisual Features

Thumbnail
Auteur
Psallidas T., Spyrou E., Perantonis S.J.
Date
2022
Language
en
DOI
10.1109/SMAP56125.2022.9941864
Sujet
Classification (of information)
Computer vision
Deep learning
Image segmentation
Video recording
Audio data
Audio-visual content
Audio-visual features
Content representation
User-generated
User-generated video
Video segments
Video summaries
Video summarization
Visual data
Large dataset
Institute of Electrical and Electronics Engineers Inc.
Afficher la notice complète
Résumé
The ever-increasing amount of user-generated audiovisual content has increased the demand for easy navigation across content collections and repositories, necessitating detailed, yet concise content representations. A typical method to this goal is to construct a visual summary, which is significantly more expressive than other alternatives, such as verbal annotations. In this paper, we describe a video summarization technique which is based on the extraction and the fusion of audio and visual data, in order to generate dynamic video summaries, i.e., video summaries that include the most essential video segments from the original video, while maintaining their original temporal sequence. Based on the extracted features, each video segment is classified as being "interesting"or "uninteresting,"and hence included or excluded from the final summary. The originality of our technique is that prior to classification, we employ a transfer learning strategy to extract deep features from pre-trained models as input to the classifiers, making them more intuitive and robust to objectiveness. We evaluate our technique on a large dataset of user-generated videos and demonstrate that the addition of deep features is able to improve classification performance, resulting in more concrete video summaries, compared to the use of only hand-crafted features. © 2022 IEEE.
URI
http://hdl.handle.net/11615/78394
Collections
  • Δημοσιεύσεις σε περιοδικά, συνέδρια, κεφάλαια βιβλίων κλπ. [19735]

Related items

Showing items related by title, author, creator and subject.

  • Thumbnail

    Κατάτμηση σε Tiles και αποδοτικότητα κωδικοποίησης video: οι περιπτώσεις των προτύπων HEVC και AV1 

    Πανάγου, Ναταλία (2019)
  • Thumbnail

    Υλοποίηση του avs video standard σε έναν massively parallel πολυεπεξεργαστή 

    Παπαπέτρου-Λαμπράκη, Νεφέλη Α. (2010)
  • Thumbnail

    Μελέτη του προτύπου video με την ονομασία "Scalable video coding" 

    Τσουμπλέκας, Γεώργιος (2010)
htmlmap 

 

Parcourir

Tout DSpaceCommunautés & CollectionsPar date de publicationAuteursTitresSujetsCette collectionPar date de publicationAuteursTitresSujets

Mon compte

Ouvrir une sessionS'inscrire
Help Contact
DepositionAboutHelpContactez-nous
Choose LanguageTout DSpace
EnglishΕλληνικά
htmlmap