Stage-oe-small.jpg

Inproceedings3603: Unterschied zwischen den Versionen

Aus Aifbportal
Wechseln zu:Navigation, Suche
 
(2 dazwischenliegende Versionen von einem anderen Benutzer werden nicht angezeigt)
Zeile 19: Zeile 19:
 
|Title=Knowledge Guided Attention and Inference for Describing Images Containing Unseen Objects
 
|Title=Knowledge Guided Attention and Inference for Describing Images Containing Unseen Objects
 
|Year=2018
 
|Year=2018
 +
|Month=Juni
 
|Booktitle=The Semantic Web. Latest Advances and New Domains. 15th Extended Semantic Web Conference (ESWC), Crete, Greece.
 
|Booktitle=The Semantic Web. Latest Advances and New Domains. 15th Extended Semantic Web Conference (ESWC), Crete, Greece.
 
|Publisher=Springer International Publishing.
 
|Publisher=Springer International Publishing.
 
}}
 
}}
 
{{Publikation Details
 
{{Publikation Details
|Abstract=Images on the Web encapsulate diverse knowledge about var-
+
|Abstract=Images on the Web encapsulate diverse knowledge about varied abstract concepts. They cannot be sufficiently described with models learned from image-caption pairs that mention only a small number of visual object categories. In contrast, large-scale knowledge graphs contain many more concepts that can be detected by image recognition models. Hence, to assist description generation for those images which contain visual objects unseen in image-caption pairs, we propose a two-step process by leveraging large-scale knowledge graphs. In the first step, a multi-entity recognition model is built to annotate images with concepts not mentioned in any caption. In the second step, those annotations are leveraged as external semantic attention and constrained inference
ied abstract concepts. They cannot be sufficiently described with models learned from image-caption pairs that mention only a small number of visual object categories. In contrast, large-scale knowledge graphs contain many more concepts that can be detected by image recognition models. Hence, to assist description generation for those images which contain visual objects unseen in image-caption pairs, we propose a two-step process by leveraging large-scale knowledge graphs. In the first step, a multi-entity recognition model is built to annotate images with concepts not mentioned in any caption. In the second step, those annotations
+
in the image description generation model. Evaluations show that our models outperform most of the prior work on out-of-domain MSCOCO image description generation and also scales better to broad domains with more unseen objects.
are leveraged as external semantic attention and constrained inference in the image description generation model. Evaluations show that our models outperform most of the prior work on out-of-domain MSCOCO image description generation and also scales better to broad domains with more unseen objects.
+
|Download=Eswc2018-camera-ready.pdf,Presentation-ImageCapUnknownObj-Rettinger-ESWC18.pdf,
 
|Forschungsgruppe=Web Science
 
|Forschungsgruppe=Web Science
 
}}
 
}}

Aktuelle Version vom 4. Juni 2018, 13:51 Uhr


Knowledge Guided Attention and Inference for Describing Images Containing Unseen Objects


Knowledge Guided Attention and Inference for Describing Images Containing Unseen Objects



Published: 2018 Juni

Buchtitel: The Semantic Web. Latest Advances and New Domains. 15th Extended Semantic Web Conference (ESWC), Crete, Greece.
Verlag: Springer International Publishing.

Nicht-referierte Veröffentlichung

BibTeX

Kurzfassung
Images on the Web encapsulate diverse knowledge about varied abstract concepts. They cannot be sufficiently described with models learned from image-caption pairs that mention only a small number of visual object categories. In contrast, large-scale knowledge graphs contain many more concepts that can be detected by image recognition models. Hence, to assist description generation for those images which contain visual objects unseen in image-caption pairs, we propose a two-step process by leveraging large-scale knowledge graphs. In the first step, a multi-entity recognition model is built to annotate images with concepts not mentioned in any caption. In the second step, those annotations are leveraged as external semantic attention and constrained inference in the image description generation model. Evaluations show that our models outperform most of the prior work on out-of-domain MSCOCO image description generation and also scales better to broad domains with more unseen objects.

Download: Media:Eswc2018-camera-ready.pdf,Media:Presentation-ImageCapUnknownObj-Rettinger-ESWC18.pdf



Forschungsgruppe

Web Science


Forschungsgebiet