• Aucun résultat trouvé

Recommendations Based on Visual Content

N/A
N/A
Protected

Academic year: 2022

Partager "Recommendations Based on Visual Content"

Copied!
3
0
0

Texte intégral

(1)

Recommendations Based on Visual Content

Taras Hnot

National University of Life and Environmental Science of Ukraine, Department of Economic Cybernetics

tarashnot@gmail.com

Abstract. There is a large number of algorithms to perform recommendations for customers of online platforms. All depends on the data sources we have. Widely used approaches are based on transactional data and “ratings” matrices. For such kind of products as clothes, furniture, hand clocks it is very important to take into account not only some metadata characteristics, but also their “visual look”. Peo- ple always buy clothes based not only on their size, sleeves lengths, textile type, etc., but based on how it looks in general. In this poster paper, we will show how feature vectors of visual content could be extracted and used to enhance recom- mendations.

Keywords: Visual Recommendations, Deep Neural Networks, ResNet50, Deep features representation, fine-tuning of NN.

1 Introduction

In today’s world, there are multiple ways to perform recommendations starting from using attributes and metadata of the products and ending with rates, received by multi- ple users. In this paper, we are proposing to include into recommendations also visual information, which could be extracted from photos of the products.

Visual information is stored in pixel values of the images. But exact pixels’ repre- sentation is not the best way to represent images’ features. These values are shifted towards position of the object on the image, lighting, etc. It is better to use some

“deeper” representation, which could be extracted using neural networks.

2 Neural Networks for Deep Features Extraction

The best way to extract features of images is to use some neural network, which uses these features to perform classification. Earlier layers of such networks give us an op- portunity to represent images in the best possible way for comparison with nearest neighbors.

Fine-tuning [3] is a frequently used approach while training neural networks with images. The main idea is to use already trained model and only slightly tune it to work with new data of the same nature. This approach is very useful while working with limited number of data. For example, to train image classifier from scratch we need

(2)

tens of thousands of observations per class and days of training to achieve high accu- racy. In case of fine-tuning it will be enough to have just few hundreds of images per class and a model could be trained in just a few minutes. This could be achieved by using pretrained deep features and building even linear classifier on top of them.

To create a model, subset of DeepFashion[2] dataset was used (46,985 images) to train 46-classes classifier (shirt, cutoff, jeans, suit, etc.). Subset of it could be seen on Fig.1.

Fig. 1. Sample of DeepFashion dataset

To do fine-tuning, some base model is needed. For that purpose we have used Res- Net5[1] trained on ImageNet (1000 classes of 1.28 millions of images).

Process of training is next:

1. Cut off last output layer with 1000 neurons;

2. Add two fully-connected layers (256, 64 neurons) with RELU activation;

3. Add output layer with 46 possible outputs with SOFTMAX activation;

4. Freeze all weight except just added;

5. Train new weights for 10 epochs with ADAM optimizer;

6. Unfreeze all other weights;

7. Fine-tune all weights for 10 epochs with very small learning rate, like (0.001).

Following approach described above, we have achieved 0.76 top 3 accuracy (top 3 means that observation is classified correctly if true value is predicted in top 3 classi- fier’s outputs).

Then deep image features could be extracted from network using activation of layer before two last layers, which perform classification. In our case – they are 1000 numeric vectors. After the whole dataset of ~290 thousands of images was processed to extract features vectors, comparison was performed using Euclidean distance. Achieved results are on Fig.2.

(3)

Fig.2. Visual recommendations (first column – input images, next columns – visually similar, sorted by distance, starting from the closest one)

These kind of recommendations are not final. They could be improved by incorpo- rating into feature vector information, related to e.g., color, style, patter.

3 Conclusions

Deep features give an ability to extract information from a visual content that is im- portant for specific task. In our work, we have showed that models, which are used to extract these features, could be trained easily on small data sets using such technique as fine-tuning.

References

1. Kaiming He, Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016).

2. Ziwei Liu, Sijie Yan, Ping Luo, Xiaogang Wang, Xiaoou Tang: DeepFashion: Powering Robust Clothes Recognition and Retrieval with Rich Annotations. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (2016).

3. Ec Lin, Z., Ji, K., Kang, M., Leng, X., Zou, H.: Deep Convolutional Highway Unit Network for SAR Target Classification with Limited Labeled Training Data. IEEE Geosci. Remote Sens. Lett. 14, 1091–1095 (2017).

Références

Documents relatifs

Our experiments showed that it is possible to achieve good retrieval performance on a medical image collection using a CBIR approach.. We used global features, which is in contrast

3.1 Image feature errors proportional to z −1 Classical IBVS control based on point features leads to an error criteria along the optical axis that is in- versely proportional to

Our approach was able to build the preference model using RDF data asso- ciated with the driver history as a result of the following: (i) it used an easy translation method

• Run 2: Integer Features with Metric Spaces Hashing The extracted features are quantized, to use integer values instead of double values, and then indexed using metric spaces

Labelled images are considered as training data in learning strategy based on classification process.. We introduce an active learning strategy to select the most difficult images

DIMENSIONALITY REDUCTION OF VISUAL FEATURES USING SPARSE PROJECTORS FOR CONTENT-BASED IMAGE RETRIEVAL1. Romain Negrel ⋆ David Picard ⋆ Philippe-Henri Gosselin

Second, for a vocabulary size of 1000, classification rates are almost constant with respect to the number of patches, while for a vocabulary size of 5000, classification rates

The second component estimates, given a query, a score for each document and for each modality (textual and visual). Finally, the last component combines linearly the score obtained