Deutsch
 
Hilfe Datenschutzhinweis Impressum
  DetailsucheBrowse

Datensatz

 
 
DownloadE-Mail
  f-VAEGAN-D2: A Feature Generating Framework for Any-Shot Learning

Xian, Y., Sharma, S., Schiele, B., & Akata, Z. (2019). f-VAEGAN-D2: A Feature Generating Framework for Any-Shot Learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 10275-10284). Piscataway, NJ: IEEE. doi:10.1109/CVPR.2019.01052.

Item is

Basisdaten

einblenden: ausblenden:
Genre: Konferenzbeitrag
Latex : {\texttt{f-VAEGAN-D2}}: {A} Feature Generating Framework for Any-Shot Learning

Externe Referenzen

einblenden:
ausblenden:
Beschreibung:
This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright.
OA-Status:

Urheber

einblenden:
ausblenden:
 Urheber:
Xian, Yongqin1, Autor           
Sharma, Saurabh1, Autor           
Schiele, Bernt1, Autor           
Akata, Zeynep1, Autor           
Affiliations:
1Computer Vision and Machine Learning, MPI for Informatics, Max Planck Society, ou_1116547              

Inhalt

einblenden:
ausblenden:
Schlagwörter: -
 Zusammenfassung: When labeled training data is scarce, a promising data augmentation approach is to generate visual features of unknown classes using their attributes. To learn the class conditional distribution of CNN features, these models rely on pairs of image features and class attributes. Hence, they can not make use of the abundance of unlabeled data samples. In this paper, we tackle any-shot learning problems i.e. zero-shot and few-shot, in a unified feature generating framework that operates in both inductive and transductive learning settings. We develop a conditional generative model that combines the strength of VAE and GANs and in addition, via an unconditional discriminator, learns the marginal feature distribution of unlabeled images. We empirically show that our model learns highly discriminative CNN features for five datasets, i.e. CUB, SUN, AWA and ImageNet, and establish a new state-of-the-art in any-shot learning, i.e. inductive and transductive (generalized) zero- and few-shot learning settings. We also demonstrate that our learned features are interpretable: we visualize them by inverting them back to the pixel space and we explain them by generating textual arguments of why they are associated with a certain label.

Details

einblenden:
ausblenden:
Sprache(n): eng - English
 Datum: 20192019
 Publikationsstatus: Online veröffentlicht
 Seiten: -
 Ort, Verlag, Ausgabe: -
 Inhaltsverzeichnis: -
 Art der Begutachtung: -
 Identifikatoren: BibTex Citekey: XianCVPR2019a
DOI: 10.1109/CVPR.2019.01052
 Art des Abschluß: -

Veranstaltung

einblenden:
ausblenden:
Titel: 32nd IEEE Conference on Computer Vision and Pattern Recognition
Veranstaltungsort: Long Beach, CA, USA
Start-/Enddatum: 2019-06-16 - 2019-06-20

Entscheidung

einblenden:

Projektinformation

einblenden:

Quelle 1

einblenden:
ausblenden:
Titel: IEEE/CVF Conference on Computer Vision and Pattern Recognition
  Kurztitel : CVPR 2019
  Untertitel : Proceedings
Genre der Quelle: Konferenzband
 Urheber:
Affiliations:
Ort, Verlag, Ausgabe: Piscataway, NJ : IEEE
Seiten: - Band / Heft: - Artikelnummer: - Start- / Endseite: 10275 - 10284 Identifikator: ISBN: 978-1-7281-3293-8