Deutsch
 
Hilfe Datenschutzhinweis Impressum
  DetailsucheBrowse

Datensatz

DATENSATZ AKTIONENEXPORT
  Multi-view Priors for Learning Detectors from Sparse Viewpoint Data

Pepik, B., Stark, M., Gehler, P., & Schiele, B. (2014). Multi-view Priors for Learning Detectors from Sparse Viewpoint Data. In International Conference on Learning Representations 2014 (pp. 1-13). Ithaca, NY: Cornell University. Retrieved from http://arxiv.org/abs/1312.6095.

Item is

Dateien

einblenden: Dateien
ausblenden: Dateien
:
arXiv:1312.6095.pdf (Preprint), 4KB
Name:
arXiv:1312.6095.pdf
Beschreibung:
File downloaded from arXiv at 2014-05-22 11:57
OA-Status:
Sichtbarkeit:
Öffentlich
MIME-Typ / Prüfsumme:
application/pdf / [MD5]
Technische Metadaten:
Copyright Datum:
-
Copyright Info:
-

Externe Referenzen

einblenden:

Urheber

einblenden:
ausblenden:
 Urheber:
Pepik, Bojan1, Autor           
Stark, Michael1, Autor           
Gehler, Peter2, Autor           
Schiele, Bernt1, Autor           
Affiliations:
1Computer Vision and Multimodal Computing, MPI for Informatics, Max Planck Society, ou_1116547              
2External Organizations, ou_persistent22              

Inhalt

einblenden:
ausblenden:
Schlagwörter: Computer Science, Computer Vision and Pattern Recognition, cs.CV
 Zusammenfassung: While the majority of today's object class models provide only 2D bounding boxes, far richer output hypotheses are desirable including viewpoint, fine-grained category, and 3D geometry estimate. However, models trained to provide richer output require larger amounts of training data, preferably well covering the relevant aspects such as viewpoint and fine-grained categories. In this paper, we address this issue from the perspective of transfer learning, and design an object class model that explicitly leverages correlations between visual features. Specifically, our model represents prior distributions over permissible multi-view detectors in a parametric way -- the priors are learned once from training data of a source object class, and can later be used to facilitate the learning of a detector for a target class. As we show in our experiments, this transfer is not only beneficial for detectors based on basic-level category representations, but also enables the robust learning of detectors that represent classes at finer levels of granularity, where training data is typically even scarcer and more unbalanced. As a result, we report largely improved performance in simultaneous 2D object localization and viewpoint estimation on a recent dataset of challenging street scenes.

Details

einblenden:
ausblenden:
Sprache(n): eng - English
 Datum: 2013-12-202014-02-162014
 Publikationsstatus: Online veröffentlicht
 Seiten: 13 p., 7 figures, 4 tables, International Conference on Learning Representations 2014
 Ort, Verlag, Ausgabe: -
 Inhaltsverzeichnis: -
 Art der Begutachtung: -
 Identifikatoren: arXiv: 1312.6095
BibTex Citekey: 844
URI: http://arxiv.org/abs/1312.6095
 Art des Abschluß: -

Veranstaltung

einblenden:
ausblenden:
Titel: International Conference on Learning Representations 2014
Veranstaltungsort: Banff, Canada
Start-/Enddatum: 2014-04-14 - 2014-04-16

Entscheidung

einblenden:

Projektinformation

einblenden:

Quelle 1

einblenden:
ausblenden:
Titel: International Conference on Learning Representations 2014
  Kurztitel : ICLR 2014
Genre der Quelle: Konferenzband
 Urheber:
Affiliations:
Ort, Verlag, Ausgabe: Ithaca, NY : Cornell University
Seiten: - Band / Heft: - Artikelnummer: - Start- / Endseite: 1 - 13 Identifikator: -