Sie befinden Sich nicht im Netzwerk der Universität Paderborn. Der Zugriff auf elektronische Ressourcen ist gegebenenfalls nur via VPN oder Shibboleth (DFN-AAI) möglich. mehr Informationen...
Ergebnis 14 von 28
Excitation Backprop for RNNs
2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, p.1440-1449
2018
Volltextzugriff (PDF)

Details

Autor(en) / Beteiligte
Titel
Excitation Backprop for RNNs
Ist Teil von
  • 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, p.1440-1449
Ort / Verlag
IEEE
Erscheinungsjahr
2018
Quelle
IEEE/IET Electronic Library (IEL)
Beschreibungen/Notizen
  • Deep models are state-of-the-art for many vision tasks including video action recognition and video captioning. Models are trained to caption or classify activity in videos, but little is known about the evidence used to make such decisions. Grounding decisions made by deep networks has been studied in spatial visual content, giving more insight into model predictions for images. However, such studies are relatively lacking for models of spatiotemporal visual content - videos. In this work, we devise a formulation that simultaneously grounds evidence in space and time, in a single pass, using top-down saliency. We visualize the spatiotemporal cues that contribute to a deep model's classification/captioning output using the model's internal representation. Based on these spatiotemporal cues, we are able to localize segments within a video that correspond with a specific action, or phrase from a caption, without explicitly optimizing/training for these tasks.
Sprache
Englisch
Identifikatoren
eISSN: 2575-7075
DOI: 10.1109/CVPR.2018.00156
Titel-ID: cdi_ieee_primary_8578254

Weiterführende Literatur

Empfehlungen zum selben Thema automatisch vorgeschlagen von bX