Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/109133
Citations
Scopus Web of Science® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorYu, L.-
dc.contributor.authorYang, Y.-
dc.contributor.authorHuang, Z.-
dc.contributor.authorWang, P.-
dc.contributor.authorSong, J.-
dc.contributor.authorShen, H.T.-
dc.date.issued2016-
dc.identifier.citationIEEE Transactions on Image Processing, 2016; 25(12):5689-5701-
dc.identifier.issn1057-7149-
dc.identifier.issn1941-0042-
dc.identifier.urihttp://hdl.handle.net/2440/109133-
dc.description.abstractIn recent years, the task of event recognition from videos has attracted increasing interest in multimedia area. While most of the existing research was mainly focused on exploring visual cues to handle relatively small-granular events, it is difficult to directly analyze video content without any prior knowledge. Therefore, synthesizing both the visual and semantic analysis is a natural way for video event understanding. In this paper, we study the problem of Web video event recognition, where Web videos often describe large-granular events and carry limited textual information. Key challenges include how to accurately represent event semantics from incomplete textual information and how to effectively explore the correlation between visual and textual cues for video event understanding. We propose a novel framework to perform complex event recognition from Web videos. In order to compensate the insufficient expressive power of visual cues, we construct an event knowledge base by deeply mining semantic information from ubiquitous Web documents. This event knowledge base is capable of describing each event with comprehensive semantics. By utilizing this base, the textual cues for a video can be significantly enriched. Furthermore, we introduce a two-view adaptive regression model, which explores the intrinsic correlation between the visual and textual cues of the videos to learn reliable classifiers. Extensive experiments on two real-world video data sets show the effectiveness of our proposed framework and prove that the event knowledge base indeed helps improve the performance of Web video event recognition.-
dc.description.statementofresponsibilityLitao Yu, Yang Yang, Zi Huang-
dc.language.isoen-
dc.publisherIEEE-
dc.rights© 2016 IEEE-
dc.source.urihttp://dx.doi.org/10.1109/tip.2016.2614136-
dc.subjectVideo event recognition; event knowledge base; two-view adaptive regression-
dc.titleWeb video event recognition by semantic analysis from ubiquitous documents-
dc.typeJournal article-
dc.identifier.doi10.1109/TIP.2016.2614136-
dc.relation.granthttp://purl.org/au-research/grants/arc/DP130103252-
dc.relation.granthttp://purl.org/au-research/grants/arc/FT120100718-
dc.relation.grantZYGX2014Z007-
dc.relation.grantZYGX2015J055-
pubs.publication-statusPublished-
Appears in Collections:Aurora harvest 8
Electrical and Electronic Engineering publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.