Skip to main content
Erschienen in: Complex & Intelligent Systems 6/2022

Open Access 13.05.2022 | Survey and State of the Art

Survey on clothing image retrieval with cross-domain

verfasst von: Chen Ning, Yang Di, Li Menglu

Erschienen in: Complex & Intelligent Systems | Ausgabe 6/2022

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

The paper summarizes the research progress on critical region recognition and deep metric learning to achieve accurate clothing image retrieval in cross-domain situations. Critical region recognition is of great value for the clothing feature extraction, effectively improving retrieval accuracy. The accuracy will decrease when solving difficult samples with similar features but different categories. Nowadays, deep metric learning is an effective way to solve this problem, which utilizes the optimization of different loss functions and ensemble network to strengthen the discrimination of clothing features. Therefore, through comparison of the experimental results of different algorithms and analysis of the accuracy of cross-domain clothing retrieval, it is demonstrated that the improvement of the retrieval accuracy in the future mainly depends on clothing important feature extraction and clothing feature discrimination.
Hinweise

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Introduction

The clothing image retrieval technology is that the computer recognizes the given clothing image and recommends clothing images with similar styles. Clothing image retrieval has been widely used in e-commerce platforms and search fields, such as Taobao, Jingdong, Baidu, Google, etc. benefit from clothing image retrieval. People prefer to take photos in daily life and find their favorite clothing from the Internet. Cross-domain clothing retrieval technology can help people quickly and accurately find similar styles of clothing on the Internet. This not only meets the needs of people’s daily life and improves the quality of life, but also promotes the consumption of the clothing industry. According to the fashion industry survey, the scale of domestic and foreign fashion markets is growing steadily. By 2024, the size of the domestic fashion market is expected to increase by 8.8% compared with the size of 2020, thus the market size will reach 26.288 million US dollars [1, 2]
Although clothing image retrieval technology has made great progress in the past 10 years, clothing image retrieval in cross-domain situations still faces great challenges. Cross-domain clothing retrieval means that the image to be queried and the image retrieval database come from two different In the scene domain, clothing images in online shopping malls are retrieved from daily street clothing images based on their similarity. Referring to recent surveys, it is found that it has two main difficulties: (1) clothes are flexible items, and when viewed from different shooting angles or The appearance can be very different when wearing different body types. The query image provided by the user may be taken under complex conditions, with complex background, various shooting angles, various lighting conditions, and even occlusion. Instead, most store images feature clean backgrounds, good lighting, and good frontal angles. (2) The intra-class variance is large and the inter-class variance is small, which is an inherent characteristic of clothing images. For example, two dresses from different categories are very similar in color and design, but have subtle differences in the shape of the neckline, one is V-shaped and the other is U-shaped. Given an image of a user’s clothing with a V-neck, returning a dress with a U-neck is not considered a correct search result in the retrieval system.

Motivation

Clothing image retrieval in cross-domain situations has been widely used in daily life. In online shopping, when a user provides a photo of daily life, the retrieval system can return clothing images with the same or similar characteristics, which reduces the retrieval system’s dependence on text, so that the desired clothing images can be retrieved more directly and accurately. Cross-domain clothing image retrieval is also widely used in physical stores. To accurately avoid purchasing and avoiding inventory, store owners need to understand the clothing preferences of people in nearby blocks. The traditional algorithm is to manually count and classify the clothing styles of consumers around the store. However, if the program can automatically take pictures of nearby pedestrians and analyze the attributes of the clothes, the number of observers can be greatly increased, and the time and labor costs can be reduced at the same time, providing a stronger reference and basis for purchasing decisions. Therefore, the research and summary of cross-domain clothing image retrieval has far-reaching significance for both individuals and society.
In actual demand, in the major domestic e-commerce shopping platforms, clothing images are mainly retrieved through keywords or texts, and the essence is to search for pictures by text. This technique requires that clothing images be finely classified and labeled accordingly. But with the explosive growth of clothing images, the shortcomings of this method become more and more obvious. First, keywords can only describe the easy-to-extract and abstract semantic features, and cannot fully reflect the visual features of clothing images, especially some fine and difficult-to-describe features; secondly, due to the huge number of clothing images, it takes a lot of Human and material resources are used for manual labeling, and manual labeling is prone to subjective bias; finally, if the description of the search keywords entered by the user is not accurate enough, it is difficult to retrieve the desired product. Therefore, this paper studies the content-based cross-domain clothing image retrieval, summarizes and evaluates it from different technical perspectives, and hopes to bring some inspiration to researchers and find new research hotspots for future research.

Previous work

With the development of deep learning [36],the framework of cross-domain clothing image retrieval is shown in Fig. 1. Cross-domain clothing image retrieval mainly includes two key steps: feature extraction and similarity measurement. For clothing image features, the methods of critical region recognition is generally used to identify important areas of clothing. The new methods of similarity measurement have also appeared. At present, the best effect is use the deep metric learning methods
Table 1
The comparison of cross-domain clothing image retrieval methods
Methods
Detailed methods
Label information
Accuracy
Computation complexity
Critical region recognition
Bounding box method
Less
Lower
Low
 
Human body landmarks recognition
Less
Lower
Low
 
Clothing landmarks recognition
More
Higher
Higher
 
Attention map recognition
Less
Higher
Higher
Deep metric learning
Siamese network
No
Low
High
 
Triplet network and Variants
No
High
Higher
 
Ensemble network
No
Higher
Highest
As shown in Table 1, different methods from past studies are summarized based on the latest research survey. In cross-domain clothing retrieval, deformation, occlusion, complex background and other phenomena will occur, which bring challenges to the accuracy of cross-domain clothing retrieval. At present, most of the critical region recognition of clothing is detect foreground objects before extracting features. The main purpose is to suppress background differences, enhance the identification of relevant local details, and provide more discriminative features in image feature extraction, which is convenient for distinguishing different types of objects. Another major challenge of cross-domain clothing image retrieval is distinguish similar images of different categories, and to cluster the images with larger differences in the same category. Deep metric learning maps images to feature vectors in space through deep neural networks. In this space, Euclidean distance or cosine distance can be directly used as a distance metric between two points. The contribution of many deep metric learning algorithms design a loss function that can learn more discriminative features. Therefore, there is a large amount of work researching deep metric learning and its related loss functions, including contrastive loss, triplet loss, triplet loss of more complex mutations, and the combination of multiple networks or methods. The ensemble method of output grouping together.

Critical region recognition learning

Bounding box method

The bounding box method uses some detection methods to identify the clothing regions in the image, and uses a rectangular box to mark the frame, as shown in Fig. 2. The purpose reduces the clothing image from the complex background and other external environmental factors during the retrieval process. Enhancing the effect of neural network on the feature extraction of clothing images.
Kiapour et al. [7] used a selective search method [8] to filter out any image with a width less than one-fifth of the image width, and directly used manual labeling of the bounding box of the clothing to limit the influence of the background regions and obtain a more accurate search. And other steps help reduce some of the variability observed in different online stores and item descriptions. Chen et al. [9] based on the R-CNN [10] target detection method made some improvements to the clothing detection problem in the image, using selective search to generate expected region suggestions, and using the Network-in-Network (NIN) model feature extraction was performed on the local regions. Then, Huang et al. [11] embedded additional semantic information in the tree structure layer of the attribute perception network, after they obtained the depth features of attribute perception, use Support Vector Regression (SVR) to predict the overlap ratio of each candidate frame, limit the size range and aspect ratio of the bounding box, and discard some inappropriate candidate solutions, thereby enhancing the positioning effect of the clothing regions bounding box in the image. In general, with the development of target detection [12], bounding box method is relatively easy to implement, the speed and accuracy of the recognition are also improved. However, when solving more complicated clothing images with background, human posture, and occlusion, the image features extracted in the clothing regions set by the bounding box will have many interference features, so the accuracy of retrieval accuracy is decreased.

Human body landmark recognition method

The landmark recognition method of the human body focuses on the limbs of the person wearing clothes. As shown in Fig. 3, according to the important body parts, it identifies the important regions of clothing, and then uses convolutional neural network to perform feature analysis on this region.
A method of research work is based on the human body pose estimation method proposed by Marcin Eichner’s team [13]to realize the detection of human body important nodes, they associated the clothing regions with the important parts of the human body and divided them into 9 parts: the torso, the left and right upper arms, the left and right lower arms, the upper left and right legs, and the lower left and right legs [14, 15]. In the recognition process, the upper body detection [16] and face detection technology [17] are combined to estimate the regions of the upper body. On this basis, the human body is segmented using the GrabCut algorithm [18], and finally combined with the human body proposed in the literature [13]. The appearance estimation model estimates the posture of the human body, thereby further dividing the important regions. The human body important node recognition method uses the landmarks of the human body to realize the detection of the important parts of the clothing. When the posture of the human body in the image is complex, it can still detect the clothing features of the relevant parts. However, when some regions are blocked, the expressive power of this method is limited.

Clothing landmark recognition method

Clothing landmark recognition directly detects the landmark marked by the clothing itself, and it is a newer method to realize the location of important regions of clothing, as shown in Fig. 4.
The DeepFashion database proposed by Liu et al. [19] defines a set of clothing landmarks, corresponding to a set of landmarks on the clothing structure. For example, a set of landmarks on the upper body is positioned as left/right collar ends and left/right cuffs. The landmarks of the lower end, lower left/right lower end, lower body clothing and full body clothing are also defined, because some landmarks often included in the image are occluded, so the visibility of each set of landmarks is also marked. On this dataset, Deep Fashion Alignment (DFA) network [20] is proposed to detect landmarks. DFA consists of three stages. In each stage, the output of the previous stage is used as input, and the network uses VGG-16 as the skeleton. In the first stage, DFA uses the original image as input to predict rough landmark locations and pseudo-labels. The pseudo-labels represent landmark such as clothing category and posture; In the second stage, the output of the entire network needs to predict the offset of the landmark, and the pseudo-label represents the offset of the local landmark; the third stage uses two CNNs as branches, and each branch has the same input and output. The choice of branch is determined by the pseudo-label of Phase 2. DeepFashion also has its own shortcomings. For example, each image has only one piece of clothing and each clothing category has only 4 to 8 landmark. To solve this problem, the DeepFashion2 [21] database was proposed, which has more landmark labeling and annotation information, the Match R-CNN model is proposed on this database, which is mainly composed of three parts: Feature Network (FN), Perception Network (PN) and Matching Network (MN) composition. After the query image passes through the FN network, it is input to the PN network, and then the landmark positions are obtained through the convolutional layer and the deconvolutional layer, and finally combined with the MN network for clothing retrieval. These two large clothing databases provide great support for future clothing retrieval research. The clothing landmark recognition method is also more capable of processing clothing deformation, occlusion and details, and the retrieval accuracy is greatly improved, but a lot of labeling and annotation information, and the professional knowledge of the clothing industry is very demanding.

Attention map recognition method

The attention map recognition method uses the idea of the attention mechanism to extract the image features of the saliency or visual attention regions from the original image, and usually needs to combine the attribute information of the clothing to complete the clothing retrieval task, as shown in Fig. 5.
Clothing attributes usually refer to the semantic attributes of objects or scenes shared across categories. The attributes of clothing include color, texture, fabric and style, so attributes can be used as potential and interpretable connections between image content and abstract labels. By constructing a latent space between fine-grained labels and low-level features, it helps models find inter-class and intra-class correlations between clothing categories. The attention map recognition method is an algorithm model that does not require a large number of human-labeled bounding boxes and feature extraction of landmark. Extract effective image representation from the spatial location of the salient regions reduce the cost of annotations while ensuring the effect of clothing image retrieval.
Most of the popular attention map recognition algorithms use image attributes as external information to locate the attention of the image in the database, and use the database image as the context to infer the attention of the query image. The attention model ignores the noise background and extracts discriminative features for retrieval. Wang et al. [22] proposed a deep convolutional neural network system, TagCtxYNet, which includes a convolutional layer for image feature extraction and an attention layer for spatial attention modeling. It extracts an effective representation of the image by learning attention weights. Gu et al. [23] proposed an autonomous learning Visual Attention Model (VAM) to extract attention maps from clothing images. It includes two branch networks, one is a global branch network based on CNN, which is used to extract attention maps from clothing images. Extract the bottom layer features of the image to get the image feature map. The other is the introduction of the attention branch of the Full Convolutional Network [24] (FCN), which is used to predict the saliency regions of the image to obtain the image attention map. The Impdrop module is used to connect two branches to obtain the attention feature map, and the module introduces randomness between the attention map and the feature map. The addition of this randomness can reduce the risk of overfitting and allow the neural network to learn to more robust features, so the robustness of the model is improved. Zheng et al. [25] proposed an Attention-Based Region Transfer (ART) module to highlight the importance of foreground, which works in a rough way that is not classified. The attention mechanism in the advanced features is used to extract the foreground objects of interest and mark them when the feature distribution is aligned. Through multi-layer adversarial learning, the use of complex detection models can achieve effective cross-domain retrieval.
Attribute learning models usually treat attribute prediction as a multi-label classification problem, and treat each attribute as a category. In fact, the clothing images trained by each model are associated with a series of attributes, such as “silk pocket shirts”. However, traditional attribute learning models ignore sequence information. Although [22, 23] use the combination of the attention feature map and the image feature map to find a more effective feature representation, which improved the clothing retrieval effect. However, there is a lack of more local information and research on the contextual connection of different parts of clothing. Luo et al. [26] proposed an attention-based learning strategy in the clothing image retrieval task. By integrating global information and local information, the features of clothing images can be intuitively extracted, because these two kinds of information provide complementary mechanisms. It can describe clothing images accurately, and uses the Long Short-Term Memory (LSTM) mechanism [27] to simulate the top-down spatial relationship of different parts of clothing to obtain more discriminative feature representations. Luo et al. [28] proposed a Deep Multi-task Cross-domain Hashing (DMCH) method to jointly establish the sequence correlation between clothing attributes, and learn the attention and perception vision of clothing images features to further enhance the effect of cross-domain clothing image retrieval.

Deep metric learning

Siamese network

Chopra et al. [29] first applied the contrastive loss function to the Siamese network based on deep neural network. Kiapour et al. [7] used Siamese network to predict whether two features represent the same category. Sean Bell et al. [30] used the traditional contrastive loss function to design an end-to-end Siamese network for modeling learning. Huang et al. [11] proposed a Dual-attribute perceptual Ranking Network (DARN) for feature learning based on the Siamese network. All in all, the contrastive loss of the Siamese network is the most widely used pair of losses in calibration learning.
As shown in Fig. 6, the Siamese network architecture has two parallel feature networks, followed by a normalization operation (L2) and a contrastive loss. Jia et al. [31] defined the contrastive loss function in the literature as shown in Eq. (1).
$$\begin{aligned} \mathrm{Loss}(f(a_i),f(b_i))= & {} (1-y)\mathrm{max}\left\{ 0,m-D(f(a_i),f(b_i))\right\} ^{2} \nonumber \\&+yD(f(a_i),f(b_i))^2. \end{aligned}$$
(1)
Among f() is an embedding function that maps an image to a feature vector.y is a label, D() represents the distance between two feature vectors. The margin parameter m forces the distance between images of different categories to increase, which has a certain effect on learning ordering. On this basis, Xiong et al. [32] proposed a contrastive loss function with a bilateral distance parameter, as shown in Eq. (2).
$$\begin{aligned} \mathrm{Loss}(f(a_i),f(b_i))= & {} (1-y)\mathrm{max}\left\{ 0,m_1-D(f(a_i),f(b_i))\right\} ^{2}\nonumber \\&+y \mathrm{max}\left\{ 0,D(f(a_i),f(b_i))-m_2 \right\} ^{2}.\nonumber \\ \end{aligned}$$
(2)
Here, if the positive margin (PM) parameter is equal to the negative margin (NM), it is called symmetric double margin, otherwise we call it asymmetric double margin. The existence of positive margins makes clothing images diversified to a certain extent, which is more reasonable than forcing them to be exactly the same. Wang et al. [33] optimized the contrastive loss by adding penalty constraints, and proposed a robust contrastive loss function to improve the generalization ability of the learning network.

Triplet network and variants

The triplet loss [34] function is widely used in the triplet network model, and has achieved better results in cross-domain clothing image retrieval. The structure of the triplet network is shown in Fig. 7. The main structure is three parallel feature networks that map images into feature vectors, then normalize the feature vector, and input it into the three sets of loss functions. The triplet loss makes the distance between different clothing images larger and the distance between the same clothing images smaller.
Different from the contrastive loss function that considers the absolute distance of the pair, the triplet loss calculates the relative distance between the positive pair and the negative pair of the same reference sample, and the specific definition is shown in Eq. (3).
$$\begin{aligned}&\mathrm{Loss}(f(a_i),f(p_i),f(n_i)) \nonumber \\&\quad =\mathrm{max}\left\{ 0,m+D(f(a_i),f(p_i))^2\right. \left. -D(f(a_i),f(n_i))^2 \right\} .\nonumber \\ \end{aligned}$$
(3)
Among \(a_i\),\(p_i\) and \(n_i\), respectively, represent the reference sample, the positive sample, and the negative sample. The \(a_i\) and \(n_i\) label of is the same, the \(a_i\) and \(p_i\) label is different. m is the margin between the positive and negative pairs.
Due to the triplet contains reference samples, positive samples and negative samples, so N images can generate \(\mathcal {O}(N^3)\) sample, even if it is a medium number of images, it is impossible to consider all samples, because not all samples can provide the same information for training a model. Randomly selects triples is a very inefficient training deep embedding network, which has inspired a lot of recent work to mine difficult samples for training. Wang et al. [35] randomly selected samples as triples in the first 10 rounds of training, and dig out difficult triples in each small batch after 10 rounds of training. Cui et al. [36] used manual methods to mark difficult negative images from images with high confidence scores assigned to them in each round. Simo-Serra et al. [37] analyzed the impact of difficult positive sample mining and difficult negative sample mining, and found that the combination of positive sample mining and negative sample mining improved the discrimination ability. Song et al. [38] designed a small batch of triple loss that considers all possible three-group associations in the small batch. Liu et al. [39] proposed a cluster-level triple loss, which considered the correlation between the cluster center, the positive sample and the nearest negative sample. Ge et al. [40] introduced Hierarchical Triplet Loss (HTL) to solve the random sampling problem in the triplet training process. These studies solve the problem of how to mine difficult samples in training. They use more complex triples for training, which can not only speed up the convergence speed of the learning algorithm, but also can use the positive and negative samples of the given reference learn clearer margins and better improve the global structure embedded in the cyberspace.
However, the methods based on difficult sample mining aims to find those triples that are difficult to find in the existing network from the existing training samples. It is essentially a greedy algorithm, which makes the trained feature embedding network vulnerable to bad local optima [41]. Therefore, Zhao et al. [42] seek a method that can intentionally generate difficult triples to optimize the overall network of the network, instead of using a greedy strategy to explore existing samples only for the current network. As shown in Fig. 8, to generate the goal of difficult triples, a Hard Triplet Generation (HTG) network algorithm is proposed to optimize the network’s ability to distinguish similar samples of different categories and group different samples of the same category.
Chopra et al. [43] proposed a novel Grid Search Network (GSN) to learn feature embedding for clothing retrieval. Similar to the triplet network variant, this method assumes that the training process is a search problem, and it finds a match of reference sample images in a grid containing positive and negative images. The framework also uses reinforcement learning-based strategies to learn special feature vector conversion functions, instead of simply connecting feature vectors. When applied to feature embedding networks, it further improves the clothing image retrieval accuracy. Kuang et al. [44] proposed a Graph Reasoning Network (GRNet), the similarity pyramid network, which uses global similarity to learn and query the similarity between clothing images and clothing databases.

Ensemble network

Ensemble is a widely used method of training multiple learners to obtain a combined model, and its performance is better than a single model [45, 46]. For deep metric learning, the ensemble network connects the feature embedding learned by multiple learners. Under the constraint of the distance between a given image pair, a better embedding space can usually be obtained. A better ensemble network depends on the high performance of individual learners and the diversity between learners. However, in deep metric learning, there is not much research on the optimal architecture to generate feature embedding diversity.
According to the above-mentioned use of Siamese network or triplet network for deep metric learning, they have achieved better results. Although the image of clothing of the same category is closer, and clothing of different categories is further away from the image. However, it is difficult to directly optimize the target because of the size of the sample. Therefore, difficult sample mining is widely used to solve this problem, and it costs expensive calculations on a subset of samples considered to be difficult. However, the difficulty sample is defined relative to a specific model. Such a complicated model will treat most of the samples as easy samples, and a simple model will treat most of the samples as difficult ones, both of which are not conducive to training. Since different samples have different difficulty levels, it is difficult to define a moderately complex model, and it is also difficult to fully select difficult samples. To solve the above problems, we summarize and analyze the different methods put forward by researchers.
The above-mentioned triplet loss and variant methods are only based on a single model to mine difficult sample images, and cannot make full use of samples of different difficulty levels. Therefore, Yuan et al. [47] proposed the Hard-Aware Deeply Cascaded (HDC) Embedding model, which uses increasingly complex model sets in a cascaded manner to mine negative samples of different difficulty levels during the training process. They take advantage of the deep supervision network [48, 49],and use a contrastive loss function to train the lower layers of the network to handle easier samples, and the higher layers of the training network to handle more difficult samples. Compared with this multi-layer method, the Boosting Independent Embedding Robustly (BIER) model [50]uses an ensemble method of high-dimensional embedding method, focusing on reducing the correlation on a single layer, and dividing the high-dimensional embedding into several a learner, trained with Online Gradient Boosting (OGB). Continuous learners are trained on reweighted samples, which greatly reduces the correlation between learners, thereby reducing the correlation within the embedding and improving the robustness of the embedding. In addition, compared with the HDC model, the method allows continuous weighting of samples according to the loss function. Inspired by the BIER method, Xuan et al. [51] proposed a different method to learn robust, high-dimensional embedding spaces. Instead of reweighting the input samples to create independent output embedding.
As an important aspect of ensemble network, learners should have diversity in feature embedding. So, Kim et al. [52] proposed an Attention-based Ensemble (ABE) model, as shown in Fig. 9, (a) based on ordinary ensemble learning; (b) based on attention ensemble learning. The model ensembles multiple attention models so that each learner can pay attention to different parts of the object. For different positions of different learners, different feature embedding functions are trained. A divergence loss is proposed to regularize features to distinguish feature embedding from different learners. All in all, although the ensemble method increases the complexity of the model, it can further improve the accuracy of cross-domain clothing retrieval.
Table 2
The detailed introduction of the different clothing databases
Databases
Institutions
Details
Download
Street2Shop
University of Illinois at Urbana-Champaign
404,683 store images and 20,357 street images. 39,479 street and store pairs.11 categories
DARN
National University of Singapore
450,000 store and 90,000 street images. Each image contains 5–9 semantic attributes categories.20 categories
DeepFashion
The Chinese University of Hong Kong
800,000 images, including four benchmarks, with a large amount of clothing labeling information, 50 categories
ModaNet [53]
eBay Inc
55,000 street images, clothing is more fashionable. 13 categories. Used in clothing object detection and segmentation
FashionAI [54]
The Hong Kong Polytechnic University, Hong Kong SAR
375,000 images, 6 categories of women’s clothing, a total of 41 subcategories, with landmark annotations
DeepFashion2
The Chinese University of Hong Kong
491,000 images. 13 categories with more annotation information. A comprehensive database that can be used for clothing detection,pose estimation, segmentation and retrieval

Comparison of experimental methods

Clothing databases

As shown in Table  2, the detailed introduction to the clothing databases. In recent years, the size of popular clothing databases and types of annotations is different. For example, Street2Shop and DARN contain 425K and 540K clothing images respectively. They contain two types of images (1) street images, which are images of people actually wearing clothes under daily uncontrolled environmental conditions; (2) shop images, which are clothing images of online clothing stores, which are made by professionals shot of a more controlled environment. Since the clothing category tags are extracted from the metadata of the images collected by online shopping sites, this makes their tags appear a lot of errors and confusion. DeepFashion and. ModaNet obtains labels by manually annotating clothing categories. In addition, different types of annotations are provided with these databases. DeepFashion is a large clothing database with comprehensive annotations, with four benchmarks, among which Consumerto-shop Benchmark is a database corresponding to street images and store images, and each clothing item’s folder contains a street image and several store images, with a total of 33,881 clothing items and 239,557 clothing images. Each image has 4–8 clothings functional regions (such as “ collar ” ) and other related fashion labels. The recognitions of these fashion landmarks are shared with all clothing categories, which makes it difficult for them to capture the rich variety of clothing images. In contrast, ModaNet’s street images have a mask of a single person, but there are no landmarks. Unlike the datasets above, DeepFashion2 contains 491K images and 801K annotations of landmarks, masks, and bounding boxes and 873,000 pairs of images, which is the most comprehensive benchmark in a clothing dataset.

Experiment preparation

The hardware and software environment used in the experiment is: Intel(R) Core(TM) i5-3570 CPU @ 3.40 GHz processor, NVIDIA GeForece GTX 1070 8GB graphics card, 8 GB memory. The operating system is Ubuntu 16.04, and the programming language is Python, the deep learning framework is Pytorch.
As shown in Table 3, the datasets used in this paper are two subsets of the DeepFashion dataset, namely In-shop Clothes Retrieval Benchmark subset and Consumer-to-shop Clothes Retrieval Benchmark subset
Table 3
A detailed introduction to the experimental dataset in this paper
Datasets
Introduction
Number of training sets
Number of test sets
Consumer-to-shop
Cross-domain clothing dataset
8000
2000
In-shop
Same domain clothing dataset
8000
2000

The evaluation of clothing retrieval

There are many clothing databases, but the commonly used evaluation methods in clothing retrieval are as follows: precision, MAP, and accuracy.
1. The precision is shown in Eq. (4):
$$\begin{aligned} P=\frac{A}{B}, \end{aligned}$$
(4)
where A is the number of similar clothing in the search results, and B represents the number of search results returned. From the precision rate formula, it can be seen that the precision rate can effectively investigate the proportion of the correct return results of the retrieval model in the retrieval results in all the returned results.
2. Although the precision rate is a statistical evaluation of the proportion of correct search results, it lacks the evaluation of the location information of the search results. Therefore, the MAP value is used as the evaluation of the location information of the search results. The MAP is shown in Eq. (5):
$$\begin{aligned} \mathrm{MAP} = \frac{1}{Q}\sum _{q=1}^\mathrm{NAP}(a), \end{aligned}$$
(5)
where Q is the number of clothing images in the retrieval database, which represents the average correct rate and the change in recall rate, that is, the regions under the \(P{-}R\) curve. MAP can reflect the overall performance of the retrieval method, but lacks insight into the details of the retrieval results.
3. The accuracy of cross-domain clothing retrieval is the most commonly used evaluation criterion. The Top-k method is generally used, as shown in Eq. (6):
$$\begin{aligned} \mathrm{P}@\mathrm{K}=\frac{1}{\vert Q \vert }\sum _{q \in Q}^Q\mathrm{hit}(q,k) \end{aligned}$$
(6)
where Q is the number of clothing images in the search database, and q represents the specific clothing image to be queried. If at least one clothing image in the Top-k list matches the image q, then hit(qk) will be set to 1, otherwise it will be set to 0.

Analysis of the results of the experiment

By summarizing the clothing image retrieval models based on deep learning in recent years, they mainly solve some difficulties in cross-domain situations. Tables 4 and  5, respectively, show two different models based on clothing critical region recognition and deep metric learning two ideas. Among them, “ Y ” and “ N ” indicates whether the algorithm uses the corresponding attribute and landmark annotations. It can be seen from the table that the network model based on critical region recognition has higher requirements for clothing to attribute labeling, and supervised learning is generally used. Even if the attention mechanism is used to greatly reduce the annotations of landmark in clothing, some weakly supervised networks are proposed, but some annotations of clothing attributes are still needed. However, most of the network models based on deep metric learning do not need landmark annotations or clothing attribute annotations, because deep metric learning is based on the characteristics of the image itself, in-depth mining of different difficult samples, and strengthening the discrimination of different extracted clothing features, then furthers extract the important clothing features that are discriminative.
Table 4
Different models based on critical region recognition
Models
Year
Attribute
Landmark
WTBI
2015
Y
N
DARN
2015
Y
N
FashionNet
2016
Y
Y
VAM
2017
Y
N
TagCtxYNet
2017
Y
N
MGN &SCN
2019
Y
N
Match R-CNN
2019
Y
Y
DMCH
2019
Y
N
Table 5
Different models based on deep metric learning
Models
Year
Attribute
Landmark
Partial-sharing
2016
N
N
HDC
2017
N
N
BIER
2017
N
N
HTL
2018
N
N
HTG
2018
N
N
ABE
2018
N
N
GSN
2019
N
N
GRNet
2019
N
N
At present, most popular clothing retrieval networks are implemented based on the DeepFashion database. The DeepFashion database has two subsets, Consumerto-shop Benchmark and In-shop Benchmark. The application scenarios are cross-domain clothing image retrieval and same-domain clothing image retrieval, as shown in the figure. As shown in Fig. 10, the performance of using the critical region recognition idea in solving the cross-domain clothing retrieval problem. The figure shows that different clothing critical region recognition algorithms have a greater impact on clothing retrieval performance. Among them, clothing landmark recognition and attention map recognition have a higher accuracy in clothing retrieval. However, FashionNet and Match R-CNN networks use clothing landmark recognition to rely too much on clothing attributes and landmark annotation information in the retrieval process, while the attention maps recognition method can solve this problem better in clothing retrieval, it can still achieve better retrieval accuracy without the clothing landmark annotation, which provides new ideas of cross-domain clothing image retrieval.
As shown in Fig. 11, the performance of several deep metric learning-based clothing image retrieval models in different datasets. From the comparison of the performance of (a) and (b), We find that the deep metric learning is better in the same domain clothing image retrieval than the cross-domain clothing image retrieval, because the same domain clothing image itself is less affected by the external environment. Mainly considering the influence of the inherent attributes of clothing images, deep metric learning can use different loss functions combined with network models to achieve better similar clothing matching. However, for cross-domain clothing image retrieval, deep metric learning using contrastive loss, ternary loss and variant, and ensemble learning needs to consider the influence of background and other factors, and needs to mine difficult samples. At present, ensemble learning is widely used, and it can be very useful. It is good to mine samples of different difficulty levels to improve the accuracy of cross-domain clothing retrieval.
Clothing image retrieval requires feature extraction and similarity match. Different research methods have different focuses. Fig. 12 and Table  6 show the performance of deep network models in cross-domain clothing retrieval in recent years. It can be seen that the overall effect of deep metric learning is not as good as clothing critical region recognition of solving cross-domain retrieval problems, indicating that the main problem to be solved in cross-domain clothing retrieval are the recognition of clothing important regions in the image. This is important step in using convolutional neural networks to extract features, then combine clothing attributes can achieve better retrieval results. Therefore, in the future, we can combine critical region recognition and deep metric learning to propose a new algorithm without additional annotation information, it can achieve better cross-domain clothing image retrieval accuracy.
Table 6
The retrieval results on Consumer-to-shop benchmark
Methods
Model
MAP(%)
P(%)
Critical region recognition
WTBI
32.1
22.1
DARN
41.2
30.3
FashionNet
50.1
35.6
VAM
54.3
42.1
DMCH
62.3
46.4
Deep metric learning
HDC
45.2
28.4
HTL
52.1
35.1
ABE
54.1
42.8
GSN
55.6
43.2
GRNet
60.2
44.1

Conclusion

This paper reviews the cross-domain clothing retrieval methods. First, it analyzes the common methods of the critical region recognition and deep metric learning in cross-domain clothing retrieval. Then, the research results show that attention map recognition method not only saves time and cost, but also further improves the effect of clothing retrieval. Deep metric learning research is widely used, and has achieved better results in the same-domain clothing retrieval and cross-domain clothing retrieval. At last, we find that the critical region recognition can extract more important clothing detail features, and deep metric learning makes the extracted features more discriminative, both affect the effect of cross-domain clothing retrieval.
In summary, although cross-domain clothing retrieval has achieved better retrieval results using clothing critical region recognition and deep metric learning methods, there are still many issues need to be solved, it mainly includes:
1.
Attribute labeling problem: Most of the deep network models need the assistance of clothing attribute labels, which is a supervised learning or weakly supervised method. It requires high clothing labeling and is a time-consuming and labor-intensive work. How to reduce clothing attribute labels, save costs, and improve accuracy still needs further research.
 
2.
Model complexity problem: In recent years, the research of cross-domain clothing image retrieval tasks has mainly focused on ensemble methods. Although better results have been achieved, the long training time and memory loss brought by model ensemble methods are difficult to solve. Therefore, how to solve the high model complexity brought by ensemble learning under the premise of ensuring the retrieval effect is a big challenge.
 
3.
Clothing databases: At present, the clothing databases contain different types of clothing, which are distinguished from clothing categories, such as dresses, jeans, shirts, etc. However, due to the development of the clothing fashion industry, different clothing combinations can produce ever-changing clothing styles, such as sports, Japanese, punk, etc. The retrieval of clothing styles will be another important research direction in the future.
 

Acknowledgements

This work is supported in part by China National Textile and Apparel Council (no. 2018097). National Natural science Foundation of China (61902301). Shaanxi Provincial Education Department (19JK036418JK0334) and Science and Technology Plan Project of Shaanxi Province (2022JM-146,2022JZ-35). Thanks all reviewers.

Declarations

Conflict of interest

On behalf of all authors, the corresponding author states that there is no conflict of interest.

Funding

This work is supported in part by China National Textile and Apparel Council No. 2018097, National Natural science Foundation of China 61902301, Shaanxi Provincial Education Department 19JK036418JK0334 and Science and Technology Plan Project of Shaanxi Province 2022JM-146, Thanks all reviewers.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Literatur
1.
Zurück zum Zitat Korea Federation of Textile Industries (2019) Korea fashion market trend 2019 Report; Korea Federation of Textile Industries: Seoul, Korea Korea Federation of Textile Industries (2019) Korea fashion market trend 2019 Report; Korea Federation of Textile Industries: Seoul, Korea
2.
Zurück zum Zitat Korea Fashion Association (2019) Global fashion industry survey. Seoul, Korea, Korea Fashion Association Korea Fashion Association (2019) Global fashion industry survey. Seoul, Korea, Korea Fashion Association
3.
5.
Zurück zum Zitat Chen L, Papandreou G, Schroff F, Adam H (2017) Rethinking atrous convolution for semantic image segmentation. CoRR. arXiv:1706.05587 Chen L, Papandreou G, Schroff F, Adam H (2017) Rethinking atrous convolution for semantic image segmentation. CoRR. arXiv:​1706.​05587
6.
Zurück zum Zitat He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778 He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
7.
Zurück zum Zitat Kiapour MH, Han X, Lazebnik S (2015) Where to buy it: matching street clothing photos in online shops. IEEE Int Conf Comput Vis 2015:3343–3351 Kiapour MH, Han X, Lazebnik S (2015) Where to buy it: matching street clothing photos in online shops. IEEE Int Conf Comput Vis 2015:3343–3351
8.
Zurück zum Zitat Sande KEA van de, Uijlings JRR, Gevers T, Smeulders AWM (2011) Segmentation as selective search for object recognition. In: ICCV Sande KEA van de, Uijlings JRR, Gevers T, Smeulders AWM (2011) Segmentation as selective search for object recognition. In: ICCV
9.
Zurück zum Zitat Chen Q, Huang J, Feris R, Brown L, Dong J, Yan S (2015) Deep domain adaptation for describing people based on fine-grained clothing attributes. In: CVPR Chen Q, Huang J, Feris R, Brown L, Dong J, Yan S (2015) Deep domain adaptation for describing people based on fine-grained clothing attributes. In: CVPR
10.
Zurück zum Zitat Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: CVPR Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: CVPR
11.
Zurück zum Zitat Huang J, Feris RS, Chen Q (2015) Cross-domain image retrieval with a dual attribute-aware ranking network. In: Proceedings of 2015 international conference on computer vision, pp 1062–1070 Huang J, Feris RS, Chen Q (2015) Cross-domain image retrieval with a dual attribute-aware ranking network. In: Proceedings of 2015 international conference on computer vision, pp 1062–1070
12.
Zurück zum Zitat Ning C, Menglu L, Hao Y, Xueping S, Yunhong L (2020) Survey of pedestrian detection with occlusion. Compl Intell Syst 2020:5 Ning C, Menglu L, Hao Y, Xueping S, Yunhong L (2020) Survey of pedestrian detection with occlusion. Compl Intell Syst 2020:5
13.
Zurück zum Zitat Eichner M, Ferrar V (2012) Appearance sharing for collective human pose estimation. Comput Vis ACCV 2012:38–151 Eichner M, Ferrar V (2012) Appearance sharing for collective human pose estimation. Comput Vis ACCV 2012:38–151
14.
Zurück zum Zitat Chen H, Andrew G, Bernd G (2012) Describing clothing by semantic attributes. In: Proceedings of the 12th European conference on computer vision, pp 609–623 Chen H, Andrew G, Bernd G (2012) Describing clothing by semantic attributes. In: Proceedings of the 12th European conference on computer vision, pp 609–623
15.
Zurück zum Zitat Chen K, Luo T, Jai B (2017) When fashion meets big data: discriminative mining of best selling clothing features. In: Proceedings of the 26th international conference on world wide web companion, pp 15–22 Chen K, Luo T, Jai B (2017) When fashion meets big data: discriminative mining of best selling clothing features. In: Proceedings of the 26th international conference on world wide web companion, pp 15–22
16.
Zurück zum Zitat Pedro FF, Ross BG, David AM (2010) Object detection with discriminatively trained part-based models. IEEE Trans 2010:1627–1645 Pedro FF, Ross BG, David AM (2010) Object detection with discriminatively trained part-based models. IEEE Trans 2010:1627–1645
17.
Zurück zum Zitat Viola P, Jones M (2001) Robust real-time object detection. Int J Comput Vis 57:2 Viola P, Jones M (2001) Robust real-time object detection. Int J Comput Vis 57:2
18.
Zurück zum Zitat Carten R, Vladimir K, Aadrew B (2014) “GrabCut’’: interactive foreground extraction using iterated graph cuts. ACM Trans 23(3):309–314 Carten R, Vladimir K, Aadrew B (2014) “GrabCut’’: interactive foreground extraction using iterated graph cuts. ACM Trans 23(3):309–314
19.
Zurück zum Zitat Liu Z, Luo P, Qiu S, Wang X, Tang X (2016) Deepfashion: powering robust clothes recognition and retrieval with rich annotations. Comput Vis Pattern Recogn 2016:1096–1104 Liu Z, Luo P, Qiu S, Wang X, Tang X (2016) Deepfashion: powering robust clothes recognition and retrieval with rich annotations. Comput Vis Pattern Recogn 2016:1096–1104
20.
Zurück zum Zitat Liu Z, Yan S, Lou P (2016) Fashion landmark detection in the wild. In: proceedings of the 14th European conference of computer vision (ECCV), pp 229–245 Liu Z, Yan S, Lou P (2016) Fashion landmark detection in the wild. In: proceedings of the 14th European conference of computer vision (ECCV), pp 229–245
21.
Zurück zum Zitat Ge Y, Zhang R, Wu L (2019) DeepFashion2: a versatile benchmark for detection,pose estimation,segmentation and re-identification of clothing images. In: proceedings of the 2019 conference on computer vision and pattern recognition, pp 5337–5345 Ge Y, Zhang R, Wu L (2019) DeepFashion2: a versatile benchmark for detection,pose estimation,segmentation and re-identification of clothing images. In: proceedings of the 2019 conference on computer vision and pattern recognition, pp 5337–5345
22.
Zurück zum Zitat Ji X, Wang W, Liu MH, Yang Y (2017) Cross-domain image retrieval with attention modeling. In: Proceedings ACM on multimedia conference, ACM, pp 1654–1662 Ji X, Wang W, Liu MH, Yang Y (2017) Cross-domain image retrieval with attention modeling. In: Proceedings ACM on multimedia conference, ACM, pp 1654–1662
23.
Zurück zum Zitat Wang Z, Gu Y, Zhang Y, Zhou J, Gu X (2017) Clothing retrieval with visual attention model. IEEE Vis Commun Image Process 2017:5 Wang Z, Gu Y, Zhang Y, Zhou J, Gu X (2017) Clothing retrieval with visual attention model. IEEE Vis Commun Image Process 2017:5
24.
Zurück zum Zitat Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. IEEE Trans Pattern Anal Mach Intell 2015:640 Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. IEEE Trans Pattern Anal Mach Intell 2015:640
25.
Zurück zum Zitat Zheng Y, Huang D, Liu S, Wang Y (2020) Cross-domain object detection through coarse-to-fine feature adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13766–13775 Zheng Y, Huang D, Liu S, Wang Y (2020) Cross-domain object detection through coarse-to-fine feature adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13766–13775
26.
Zurück zum Zitat Luo Z,Yuan J, Yang J, Wen W (2019) Spatial constraint multiple granularity attention network for clothes retrieval. In: 2019 IEEE international conference on image processing (ICIP), IEEE, pp 859–863 Luo Z,Yuan J, Yang J, Wen W (2019) Spatial constraint multiple granularity attention network for clothes retrieval. In: 2019 IEEE international conference on image processing (ICIP), IEEE, pp 859–863
27.
Zurück zum Zitat Xu K, Ba J, Kiros R, Cho K, Courville AC, Salakhutdinov R, Zemel RS, Bengio Y (2015) Show, attend and tell: Neural image caption generation with visual attention. In: Proceedings of the 32nd international conference on machine learning, ICML 2015, Lille, France, pp 2048–2057 Xu K, Ba J, Kiros R, Cho K, Courville AC, Salakhutdinov R, Zemel RS, Bengio Y (2015) Show, attend and tell: Neural image caption generation with visual attention. In: Proceedings of the 32nd international conference on machine learning, ICML 2015, Lille, France, pp 2048–2057
28.
Zurück zum Zitat Luo Y, Wang Z, Huang Z, Yang Y, Lu H (2019) Snap and find: deep discrete cross-domain garment image retrieval. IEEE Trans Image Process 2019:5 Luo Y, Wang Z, Huang Z, Yang Y, Lu H (2019) Snap and find: deep discrete cross-domain garment image retrieval. IEEE Trans Image Process 2019:5
29.
Zurück zum Zitat Chopra S, Hadsell R, LeCun Y (2005) Learning a similarity metric discriminatively,with application to face verification. In: Computer vision and pattern recognition (CVPR), pp 539–546 Chopra S, Hadsell R, LeCun Y (2005) Learning a similarity metric discriminatively,with application to face verification. In: Computer vision and pattern recognition (CVPR), pp 539–546
30.
Zurück zum Zitat Bell S, Bala K (2015) Learning visual similarity for product design with convolutional neural networks. ACM Trans Graph 34(4):98CrossRef Bell S, Bala K (2015) Learning visual similarity for product design with convolutional neural networks. ACM Trans Graph 34(4):98CrossRef
31.
Zurück zum Zitat Jia Y, Shelhamer E, Donahue J, Karayev S, Long J, Girshick R, Guadarrama S, Darrell T (2014) Caffe: Convolutional architecture for fast feature embedding. In: Proceedings of the ACM international conference on multimedia, ACM, pp 675–678 Jia Y, Shelhamer E, Donahue J, Karayev S, Long J, Girshick R, Guadarrama S, Darrell T (2014) Caffe: Convolutional architecture for fast feature embedding. In: Proceedings of the ACM international conference on multimedia, ACM, pp 675–678
32.
Zurück zum Zitat Xiong Y, Liu N, Xu Z, Zhang Y (2016) A parameter partial-sharing cnn architecture for cross-domain clothing retrieval. In: Visual communications and image processing (VCIP), pp 1–4 Xiong Y, Liu N, Xu Z, Zhang Y (2016) A parameter partial-sharing cnn architecture for cross-domain clothing retrieval. In: Visual communications and image processing (VCIP), pp 1–4
33.
Zurück zum Zitat Wangxi SZ, Zhang W et al (2016) Matching user photos to online products with robust deep features. In: Proceedings of the 2016 ACM on international conference on multimedia retrieval, ACM, pp 7–14 Wangxi SZ, Zhang W et al (2016) Matching user photos to online products with robust deep features. In: Proceedings of the 2016 ACM on international conference on multimedia retrieval, ACM, pp 7–14
34.
Zurück zum Zitat Schroff F, Kalenichenko D, Philbin J (2015) Facenet: A unified embedding for face recognition and clustering. In: 2015 IEEE conference on computer vision and pattern recognition (CVPR), pp 815–823 Schroff F, Kalenichenko D, Philbin J (2015) Facenet: A unified embedding for face recognition and clustering. In: 2015 IEEE conference on computer vision and pattern recognition (CVPR), pp 815–823
35.
Zurück zum Zitat Wang X, Gupta A (2015) Unsupervised learning of visual representations using videos. In: ICCV Wang X, Gupta A (2015) Unsupervised learning of visual representations using videos. In: ICCV
36.
Zurück zum Zitat Cui Y , Zhou F, Lin Y, Belongie S (2015) Fine-grained categorization and dataset bootstrapping using deep metric learning with humans in the loop. arXiv:1512.05227 Cui Y , Zhou F, Lin Y, Belongie S (2015) Fine-grained categorization and dataset bootstrapping using deep metric learning with humans in the loop. arXiv:​1512.​05227
37.
Zurück zum Zitat Simo-Serra E, Trulls E, Ferraz L, Kokkinos I, Fua P, Moreno-Noguer F (2015) Discriminative learning of deep convolutional feature point descriptors. In: ICCV Simo-Serra E, Trulls E, Ferraz L, Kokkinos I, Fua P, Moreno-Noguer F (2015) Discriminative learning of deep convolutional feature point descriptors. In: ICCV
38.
Zurück zum Zitat Song HO, Xiang Y, Jegelka S, Savarese S (2016) Deep metric learning via lifted structured feature embedding. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 4004–4012 Song HO, Xiang Y, Jegelka S, Savarese S (2016) Deep metric learning via lifted structured feature embedding. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 4004–4012
39.
Zurück zum Zitat Liu H, Tian Y, Yang Y, Pang L, Huang T (2016) Deep relative distance learning: tell the difference between similar vehicles. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 2167–2175 Liu H, Tian Y, Yang Y, Pang L, Huang T (2016) Deep relative distance learning: tell the difference between similar vehicles. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 2167–2175
40.
Zurück zum Zitat Ge W, Huang W, Dong D, Scott MR (2018) Deep metric learning with hierarchical triplet loss. In: ECCV, pp 269–285 Ge W, Huang W, Dong D, Scott MR (2018) Deep metric learning with hierarchical triplet loss. In: ECCV, pp 269–285
41.
Zurück zum Zitat Song O, Xiang H, Jegelka Y, Savarese S (2016) S: deep metric learning via lifted structured feature embedding. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4004–4012 Song O, Xiang H, Jegelka Y, Savarese S (2016) S: deep metric learning via lifted structured feature embedding. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4004–4012
42.
Zurück zum Zitat Zhao Y, Jin Z, Qi G-J, Lu H, Hua X-S (2018) An adversarial approach to hard triplet generation. In: ECCV, pp 501–517 Zhao Y, Jin Z, Qi G-J, Lu H, Hua X-S (2018) An adversarial approach to hard triplet generation. In: ECCV, pp 501–517
43.
Zurück zum Zitat Chopra A, Sinha A, Gupta H, Sarkar M, Ayush K, Krishnamurthy B (2019) Powering robust fashion retrieval with information rich feature embeddings. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops Chopra A, Sinha A, Gupta H, Sarkar M, Ayush K, Krishnamurthy B (2019) Powering robust fashion retrieval with information rich feature embeddings. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops
44.
Zurück zum Zitat Kuang Z, Gao Y, Li G, Luo P, Chen Y, Lin L, Zhang W (2019) Fashion retrieval via graph reasoning networks on a similarity pyramid. In: The IEEE international conference on computer vision (ICCV), pp 3066–3075 Kuang Z, Gao Y, Li G, Luo P, Chen Y, Lin L, Zhang W (2019) Fashion retrieval via graph reasoning networks on a similarity pyramid. In: The IEEE international conference on computer vision (ICCV), pp 3066–3075
45.
Zurück zum Zitat Lin Z, Yang Z, Huang F, Chen J (2018) Regional maximum activations of convolutions with attention for cross-domain beauty and personal care product retrieval. In: 2018 ACM multimedia conference on multimedia conference, pp 2073–2077 Lin Z, Yang Z, Huang F, Chen J (2018) Regional maximum activations of convolutions with attention for cross-domain beauty and personal care product retrieval. In: 2018 ACM multimedia conference on multimedia conference, pp 2073–2077
46.
Zurück zum Zitat Xuan H, Souvenir R, Pless R (2018) Deep randomized ensembles for metric learning. In: The European conference on computer vision (ECCV), pp 723–734 Xuan H, Souvenir R, Pless R (2018) Deep randomized ensembles for metric learning. In: The European conference on computer vision (ECCV), pp 723–734
47.
Zurück zum Zitat Yuan Y, Yang K, Zhang C (2017) Hard-aware deeply cascaded embedding. In: The IEEE international conference on computer vision (ICCV), pp 814–823 Yuan Y, Yang K, Zhang C (2017) Hard-aware deeply cascaded embedding. In: The IEEE international conference on computer vision (ICCV), pp 814–823
48.
Zurück zum Zitat Lee C-Y, Xie S, Gallagher PW, Zhang Z, Tu Z (2015) Deeply-supervised nets. In: Proc. AISTATS Lee C-Y, Xie S, Gallagher PW, Zhang Z, Tu Z (2015) Deeply-supervised nets. In: Proc. AISTATS
49.
Zurück zum Zitat Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, anhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proc. CVPR Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, anhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proc. CVPR
50.
Zurück zum Zitat Opitz M, Waltner G, Possegger H, Bischof H (2017) Bier—boosting independent embeddings robustly. In: ICCV, pp 5189–5198 Opitz M, Waltner G, Possegger H, Bischof H (2017) Bier—boosting independent embeddings robustly. In: ICCV, pp 5189–5198
51.
Zurück zum Zitat Xuan H, Souvenir R, Pless R (2018) Deep randomized ensembles for metric learning. In: ECCV, pp 723–734 Xuan H, Souvenir R, Pless R (2018) Deep randomized ensembles for metric learning. In: ECCV, pp 723–734
52.
Zurück zum Zitat Kim W, Goyal B, Chawla K, Lee J, Kwon K (2018) Attention-based ensemble for deep metric learning. In: ECCV, pp 760–777 Kim W, Goyal B, Chawla K, Lee J, Kwon K (2018) Attention-based ensemble for deep metric learning. In: ECCV, pp 760–777
53.
Zurück zum Zitat Zheng S, Yang F, Kiapour MH, Piramuthu R (2018) Modanet: a large-scale street fashion dataset with polygon annotations. In: ACM multimedia Zheng S, Yang F, Kiapour MH, Piramuthu R (2018) Modanet: a large-scale street fashion dataset with polygon annotations. In: ACM multimedia
54.
Zurück zum Zitat Zou X, Kong X, Wong W, Wang C, Liu Y, Cao Y (2019) Fashionai: a hierarchical dataset for fashion understanding. In: CVPR workshop Zou X, Kong X, Wong W, Wang C, Liu Y, Cao Y (2019) Fashionai: a hierarchical dataset for fashion understanding. In: CVPR workshop
Metadaten
Titel
Survey on clothing image retrieval with cross-domain
verfasst von
Chen Ning
Yang Di
Li Menglu
Publikationsdatum
13.05.2022
Verlag
Springer International Publishing
Erschienen in
Complex & Intelligent Systems / Ausgabe 6/2022
Print ISSN: 2199-4536
Elektronische ISSN: 2198-6053
DOI
https://doi.org/10.1007/s40747-022-00750-5

Weitere Artikel der Ausgabe 6/2022

Complex & Intelligent Systems 6/2022 Zur Ausgabe

Premium Partner