As a pivotal technique of fashion visual analysis, fashion landmark detection has attracted extensive attention in recent years. However, prior works often ignore the importance of structural layout information for fashion landmark detection, which leads to ambiguous detection results of hard landmarks. In this paper, we propose a Layout-Aware Bidirectional Transfer Network(LBTNet) which first combines the layout features learning with a powerful human pose estimation backbone - HRNet. The LBTNet can learn the layout information by our proposed Group-wise Layout Embedded Module(GLEM) which can model the dependency among fashion landmarks on the convolutional layer and perform information passing among the adjacent landmarks. We also design a novel head structure called Bidirectional Transfer Module(BTM) to capture global semantic information of fashion landmarks through a bidirectional transmission path. Therefore, the LBTNet can accurately detect these hard landmarks (e.g. occluded landmarks and invisible landmarks). And the experimental results on two large-scale fashion datasets show that our LBTNet outperforms the state-of-the-art methods by a large margin.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.