Exploring the Semantics for Visual Relationship Detection

Wentong Liao (Corresponding Author), Cuiling Lan, Wenjun Zeng, Michael Ying Yang, Bodo Rosenhahn

Research output: Contribution to journalArticleAcademic

12 Downloads (Pure)

Abstract

Scene graph construction / visual relationship detection from an image aims to give a precise structural description of the objects (nodes) and their relationships (edges). The mutual promotion of object detection and relationship detection is important for enhancing their individual performance. In this work, we propose a new framework, called semantics guided graph relation neural network (SGRN), for effective visual relationship detection. First, to boost the object detection accuracy, we introduce a source-target class cognoscitive transformation that transforms the features of the co-occurent objects to the target object domain to refine the visual features. Similarly, source-target cognoscitive transformations are used to refine features of objects from features of relations, and vice versa. Second, to boost the relation detection accuracy, besides the visual features of the paired objects, we embed the class probability of the object and subject separately to provide high level semantic information. In addition, to reduce the search space of relationships, we design a semantics-aware relationship filter to exclude those object pairs that have no relation. We evaluate our approach on the Visual Genome dataset and it achieves the state-of-the-art performance for visual relationship detection. Additionally, Our approach also significantly improves the object detection performance (i.e. 4.2\% in mAP accuracy).
Original languageEnglish
Pages (from-to)1-13
Number of pages13
JournalarXiv.org
Publication statusPublished - 3 Apr 2019

    Fingerprint

Keywords

  • cs.CV
  • ITC-GOLD

Cite this

Liao, W., Lan, C., Zeng, W., Yang, M. Y., & Rosenhahn, B. (2019). Exploring the Semantics for Visual Relationship Detection. arXiv.org, 1-13.