University of Marburg at TRECVID 2008: High-level feature extraction

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschung

Autoren

  • Markus Mühling
  • Ralph Ewerth
  • Thilo Stadelmann
  • Bing Shi
  • Bernd Freisleben

Externe Organisationen

  • Universität Siegen
  • Philipps-Universität Marburg
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des Sammelwerks2008 TREC Video Retrieval Evaluation Notebook Papers and Slides
PublikationsstatusVeröffentlicht - 2008
Extern publiziertJa
VeranstaltungTREC Video Retrieval Evaluation, TRECVID 2008 - Gaithersburg, MD, USA / Vereinigte Staaten
Dauer: 17 Nov. 200818 Nov. 2008

Abstract

In this paper, we summarize our results for the high-level feature extraction task at TRECVID 2008. Our last year's high-level feature extraction system was based on low-level features as well as on state-of-the-art approaches for camera motion estimation, text detection, face detection and audio segmentation. This system served as a basis for our experiments this year and was extended in several ways. First, we paid attention to the fact that most of the concepts suffered from a small number of positive training samples while offering a huge number of negative ones. We tried to reduce this unbalance of positive and negative training samples by sub-sampling the negative instances. Furthermore, we increased the number of positive training samples by creating image variations. Both methods improved the detection results significantly, while the sub-sampling approach achieved our best result (8.27% mean inferred average precision). Second, we incorporated two further feature types: Hough features and audio low-level features. Finally, we supplemented our approach using cross-validation in order to improve the high level feature extraction results. On the one hand, we applied cross-validation for feature selection, on the other hand we tried to find the best sampling rate of negative instances for each concept.

ASJC Scopus Sachgebiete

Zitieren

University of Marburg at TRECVID 2008: High-level feature extraction. / Mühling, Markus; Ewerth, Ralph; Stadelmann, Thilo et al.
2008 TREC Video Retrieval Evaluation Notebook Papers and Slides. 2008.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschung

Mühling, M, Ewerth, R, Stadelmann, T, Shi, B & Freisleben, B 2008, University of Marburg at TRECVID 2008: High-level feature extraction. in 2008 TREC Video Retrieval Evaluation Notebook Papers and Slides. TREC Video Retrieval Evaluation, TRECVID 2008, Gaithersburg, MD, USA / Vereinigte Staaten, 17 Nov. 2008. <https://www-nlpir.nist.gov/projects/tvpubs/tv8.papers/marburg.pdf>
Mühling, M., Ewerth, R., Stadelmann, T., Shi, B., & Freisleben, B. (2008). University of Marburg at TRECVID 2008: High-level feature extraction. In 2008 TREC Video Retrieval Evaluation Notebook Papers and Slides https://www-nlpir.nist.gov/projects/tvpubs/tv8.papers/marburg.pdf
Mühling M, Ewerth R, Stadelmann T, Shi B, Freisleben B. University of Marburg at TRECVID 2008: High-level feature extraction. in 2008 TREC Video Retrieval Evaluation Notebook Papers and Slides. 2008
Mühling, Markus ; Ewerth, Ralph ; Stadelmann, Thilo et al. / University of Marburg at TRECVID 2008 : High-level feature extraction. 2008 TREC Video Retrieval Evaluation Notebook Papers and Slides. 2008.
Download
@inproceedings{dc4f88ae06544d00a26c786d9d311d87,
title = "University of Marburg at TRECVID 2008: High-level feature extraction",
abstract = "In this paper, we summarize our results for the high-level feature extraction task at TRECVID 2008. Our last year's high-level feature extraction system was based on low-level features as well as on state-of-the-art approaches for camera motion estimation, text detection, face detection and audio segmentation. This system served as a basis for our experiments this year and was extended in several ways. First, we paid attention to the fact that most of the concepts suffered from a small number of positive training samples while offering a huge number of negative ones. We tried to reduce this unbalance of positive and negative training samples by sub-sampling the negative instances. Furthermore, we increased the number of positive training samples by creating image variations. Both methods improved the detection results significantly, while the sub-sampling approach achieved our best result (8.27% mean inferred average precision). Second, we incorporated two further feature types: Hough features and audio low-level features. Finally, we supplemented our approach using cross-validation in order to improve the high level feature extraction results. On the one hand, we applied cross-validation for feature selection, on the other hand we tried to find the best sampling rate of negative instances for each concept.",
author = "Markus M{\"u}hling and Ralph Ewerth and Thilo Stadelmann and Bing Shi and Bernd Freisleben",
year = "2008",
language = "English",
booktitle = "2008 TREC Video Retrieval Evaluation Notebook Papers and Slides",
note = "TREC Video Retrieval Evaluation, TRECVID 2008 ; Conference date: 17-11-2008 Through 18-11-2008",

}

Download

TY - GEN

T1 - University of Marburg at TRECVID 2008

T2 - TREC Video Retrieval Evaluation, TRECVID 2008

AU - Mühling, Markus

AU - Ewerth, Ralph

AU - Stadelmann, Thilo

AU - Shi, Bing

AU - Freisleben, Bernd

PY - 2008

Y1 - 2008

N2 - In this paper, we summarize our results for the high-level feature extraction task at TRECVID 2008. Our last year's high-level feature extraction system was based on low-level features as well as on state-of-the-art approaches for camera motion estimation, text detection, face detection and audio segmentation. This system served as a basis for our experiments this year and was extended in several ways. First, we paid attention to the fact that most of the concepts suffered from a small number of positive training samples while offering a huge number of negative ones. We tried to reduce this unbalance of positive and negative training samples by sub-sampling the negative instances. Furthermore, we increased the number of positive training samples by creating image variations. Both methods improved the detection results significantly, while the sub-sampling approach achieved our best result (8.27% mean inferred average precision). Second, we incorporated two further feature types: Hough features and audio low-level features. Finally, we supplemented our approach using cross-validation in order to improve the high level feature extraction results. On the one hand, we applied cross-validation for feature selection, on the other hand we tried to find the best sampling rate of negative instances for each concept.

AB - In this paper, we summarize our results for the high-level feature extraction task at TRECVID 2008. Our last year's high-level feature extraction system was based on low-level features as well as on state-of-the-art approaches for camera motion estimation, text detection, face detection and audio segmentation. This system served as a basis for our experiments this year and was extended in several ways. First, we paid attention to the fact that most of the concepts suffered from a small number of positive training samples while offering a huge number of negative ones. We tried to reduce this unbalance of positive and negative training samples by sub-sampling the negative instances. Furthermore, we increased the number of positive training samples by creating image variations. Both methods improved the detection results significantly, while the sub-sampling approach achieved our best result (8.27% mean inferred average precision). Second, we incorporated two further feature types: Hough features and audio low-level features. Finally, we supplemented our approach using cross-validation in order to improve the high level feature extraction results. On the one hand, we applied cross-validation for feature selection, on the other hand we tried to find the best sampling rate of negative instances for each concept.

UR - http://www.scopus.com/inward/record.url?scp=84905178027&partnerID=8YFLogxK

M3 - Conference contribution

BT - 2008 TREC Video Retrieval Evaluation Notebook Papers and Slides

Y2 - 17 November 2008 through 18 November 2008

ER -