Sie befinden Sich nicht im Netzwerk der Universität Paderborn. Der Zugriff auf elektronische Ressourcen ist gegebenenfalls nur via VPN oder Shibboleth (DFN-AAI) möglich. mehr Informationen...
•Proposing a multi-layer non-local feature fusion network to extract visual features.•Learning multi-modal data to model the compatibility of paired clothing.•Feature fusion operation helps to improve the performance of the model.
Clothing plays a critical role in people’s daily lives, a perfect styling clothing is able to help people to avoid weaknesses and show their personal temperament, however, not everyone is good at styling. Compatibility is the core of styling, however, determining whether a pair of garments are compatible with each other is a challenging styling issue. Years of research have been devoted to fashion compatibility learning, whereas there are still several drawbacks in visual feature detection and compatibility calculation. In this paper, we propose an end-to-end framework to learn the compatibility among tops and bottoms. In order to improve the effects of visual feature extraction, a Multi-layer Non-local Feature Fusion framework (MNLFF) is developed. Feature fusion model is used to combine both high and low-level features, while non-local block is for global feature detection. We compare our technique with the prior state-of-the-art methods in the outfit compatibility prediction task and extensive experiments on existing datasets demonstrate its effectiveness.