Pedestrian attribute recognition has attracted extensive attention in surveillance scenarios. Many researchers treat it as a multi-label classification problem. However, they ignore the relations among attributes. For example, long hair and female often appear together at the semantic level, and there exists an overlap between their attribute-related regions at the visual level, e.g., female-related regions including long-hair-related head area and skirt-related areas. There is a correspondence between the semantic level and the visual level. However, existing relation-based methods which explore relations among attributes mostly ignore the correspondence between semantic and visual levels in relations among attributes. Our main contribution lies in proposing a novel end-to-end trainable Attribute-related Graph Visual Dependencies (AGVD) framework that can learn the dependencies of attributes based on extracted attribute-related visual features. Moreover, we propose a Graph Fusion mechanism to guide the attribute-related graph generation and pruning via semantic relations innovatively. We conduct experiments on three large-scale pedestrian attribute datasets. The experiments on PETA, RAP, and PA-100k have demonstrated that the AGVD outperforms the previous state-of-the-art methods.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.