Exploring Content-based Artwork Recommendation with ...processing systems. 1097–1105. [8] Augusto...

3
Exploring Content-based Artwork Recommendation with Metadata and Visual Features Pablo Messina Ponticia Universidad Catolica de Chile Santiago, Chile [email protected] Vicente Dominguez Ponticia Universidad Catolica de Chile Santiago, Chile [email protected] Denis Parra Ponticia Universidad Catolica de Chile Santiago, Chile [email protected] Christoph Traner Modul University Vienna, Austria [email protected] Alvaro Soto Ponticia Universidad Catolica de Chile Santiago, Chile [email protected] ABSTRACT Compared to other areas, artwork recommendation has received lile aention, despite the continuous growth of the artwork mar- ket. Previous research has relied on ratings and metadata to make artwork recommendations, as well as visual features extracted with deep neural networks (DNN). However, these features have no direct interpretation to explicit visual features (e.g. brightness, texture) which might hinder explainability and user-acceptance. In this work, we study the impact of artwork metadata as well as visual features (DNN-based and aractiveness-based) for physi- cal artwork recommendation, using images and transaction data from the UGallery online artwork store. Our results indicate that: (i) visual features perform beer than manually curated data, (ii) DNN-based visual features perform beer than aractiveness-based ones, and (iii) a hybrid approach improves the performance fur- ther. Our research can inform the development of new artwork recommenders relying on diverse content data. KEYWORDS Recommender systems, Artwork Recommendation, Visual Features, Deep Neural Networks ACM Reference format: Pablo Messina, Vicente Dominguez, Denis Parra, Christoph Traner, and Al- varo Soto. 2017. Exploring Content-based Artwork Recommendation with Metadata and Visual Features. In Proceedings of ACM Conference, Washing- ton, DC, USA, July 2017 (Conference’17), 3 pages. DOI: 10.475/123 4 1 INTRODUCTION Compared to markets aected by 2008’s nancial crisis, online artwork sales are booming due to social media and new consump- tion behavior of millennials. Online art sales reached $3.27 billions in 2015, and at the current grow rate, they will reach $9.58 bil- lion by 2020 [5]. Notably, although many online businesses utilize Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for prot or commercial advantage and that copies bear this notice and the full citation on the rst page. Copyrights for third-party components of this work must be honored. For all other uses, contact the owner/author(s). Conference’17, Washington, DC, USA © 2017 Copyright held by the owner/author(s). 123-4567-24-567/08/06. . . $00.00 DOI: 10.475/123 4 recommendation systems to boost their revenue, online artwork recommendation has received lile aention compared to other areas such as movies [1] or music [3]. Previous research has shown the potential of personalized recommendations in the arts domain, such as the CHIP project [2], that implemented a personalized rec- ommendation system for the Rijksmuseum. More recently, He et al. [6] used pre-trained deep neural networks (DNN) for recom- mendation of digital art, obtaining good results. Unfortunaly, their method is not applicable for the physical artwork problem as the method assumes that the same item can be bought over and over again. Hence their work only works under the collaborative lter- ing assumption and also did not investigate explicit visual features nor metadata. Objective. In this paper, we investigate the impact of dierent features for recommending physical artworks. In particular, we reveal the utility of artwork metadata, latent (DNN) and explicit visual features extracted from images. We address the problem of artwork recommendation with positive-only feedback (user trans- actions) over one-of-a-kind items, i.e., only one instance of each artwork (paintings) is available in the dataset. Research estions. Our work was driven by the following re- search questions: RQ1. How do manually-curated metadata perform compared to visual features?, RQ2. How do latent visual features from pre-trained DNNs and explicit visual features perform and compare to each other?, and RQ3. Do feature combinations provide the best recommendation performance? Contributions. Our work makes a contribution to the unex- plored problem of recommending physical artworks. We run simu- lated experiments with real-world transaction data provided by a popular online artwork store based in USA named UGallery 1 . We also introduce a hybrid artwork recommender which exploits all features at the same time. Our results indicate that visual features perform beer than manually-curated metadata. In addition, we show that DNN features work beer than explicit aractiveness- based visual features. 2 PROBLEM DESCRIPTION e online web store UGallery supports young and emergent artists by helping them to sell their artworks over their online plaform. 1 hp://www.UGallery.com/ arXiv:1706.05786v3 [cs.IR] 23 Oct 2017

Transcript of Exploring Content-based Artwork Recommendation with ...processing systems. 1097–1105. [8] Augusto...

Page 1: Exploring Content-based Artwork Recommendation with ...processing systems. 1097–1105. [8] Augusto Q Macedo, Leandro B Marinho, and Rodrygo LT Santos. 2015. Context-aware event recommendation

Exploring Content-based Artwork Recommendation withMetadata and Visual Features

Pablo MessinaPonti�cia Universidad Catolica de

ChileSantiago, [email protected]

Vicente DominguezPonti�cia Universidad Catolica de

ChileSantiago, Chile

[email protected]

Denis ParraPonti�cia Universidad Catolica de

ChileSantiago, Chile

[email protected]

Christoph Tra�nerModul UniversityVienna, Austria

christoph.tra�[email protected]

Alvaro SotoPonti�cia Universidad Catolica de

ChileSantiago, [email protected]

ABSTRACTCompared to other areas, artwork recommendation has receivedli�le a�ention, despite the continuous growth of the artwork mar-ket. Previous research has relied on ratings and metadata to makeartwork recommendations, as well as visual features extracted withdeep neural networks (DNN). However, these features have nodirect interpretation to explicit visual features (e.g. brightness,texture) which might hinder explainability and user-acceptance.

In this work, we study the impact of artwork metadata as wellas visual features (DNN-based and a�ractiveness-based) for physi-cal artwork recommendation, using images and transaction datafrom the UGallery online artwork store. Our results indicate that:(i) visual features perform be�er than manually curated data, (ii)DNN-based visual features perform be�er than a�ractiveness-basedones, and (iii) a hybrid approach improves the performance fur-ther. Our research can inform the development of new artworkrecommenders relying on diverse content data.

KEYWORDSRecommender systems, Artwork Recommendation, Visual Features,Deep Neural NetworksACM Reference format:Pablo Messina, Vicente Dominguez, Denis Parra, Christoph Tra�ner, and Al-varo Soto. 2017. Exploring Content-based Artwork Recommendation withMetadata and Visual Features. In Proceedings of ACM Conference, Washing-ton, DC, USA, July 2017 (Conference’17), 3 pages.DOI: 10.475/123 41 INTRODUCTIONCompared to markets a�ected by 2008’s �nancial crisis, onlineartwork sales are booming due to social media and new consump-tion behavior of millennials. Online art sales reached $3.27 billionsin 2015, and at the current grow rate, they will reach $9.58 bil-lion by 2020 [5]. Notably, although many online businesses utilize

Permission to make digital or hard copies of part or all of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor pro�t or commercial advantage and that copies bear this notice and the full citationon the �rst page. Copyrights for third-party components of this work must be honored.For all other uses, contact the owner/author(s).Conference’17, Washington, DC, USA© 2017 Copyright held by the owner/author(s). 123-4567-24-567/08/06. . .$00.00DOI: 10.475/123 4

recommendation systems to boost their revenue, online artworkrecommendation has received li�le a�ention compared to otherareas such as movies [1] or music [3]. Previous research has shownthe potential of personalized recommendations in the arts domain,such as the CHIP project [2], that implemented a personalized rec-ommendation system for the Rijksmuseum. More recently, He etal. [6] used pre-trained deep neural networks (DNN) for recom-mendation of digital art, obtaining good results. Unfortunaly, theirmethod is not applicable for the physical artwork problem as themethod assumes that the same item can be bought over and overagain. Hence their work only works under the collaborative �lter-ing assumption and also did not investigate explicit visual featuresnor metadata.

Objective. In this paper, we investigate the impact of di�erentfeatures for recommending physical artworks. In particular, wereveal the utility of artwork metadata, latent (DNN) and explicitvisual features extracted from images. We address the problem ofartwork recommendation with positive-only feedback (user trans-actions) over one-of-a-kind items, i.e., only one instance of eachartwork (paintings) is available in the dataset.

Research�estions. Our work was driven by the following re-search questions: RQ1. How domanually-curatedmetadata performcompared to visual features?, RQ2. How do latent visual featuresfrom pre-trained DNNs and explicit visual features perform andcompare to each other?, and RQ3. Do feature combinations providethe best recommendation performance?

Contributions. Our work makes a contribution to the unex-plored problem of recommending physical artworks. We run simu-lated experiments with real-world transaction data provided by apopular online artwork store based in USA named UGallery1. Wealso introduce a hybrid artwork recommender which exploits allfeatures at the same time. Our results indicate that visual featuresperform be�er than manually-curated metadata. In addition, weshow that DNN features work be�er than explicit a�ractiveness-based visual features.2 PROBLEM DESCRIPTION�e online web store UGallery supports young and emergent artistsby helping them to sell their artworks over their online pla�form.

1h�p://www.UGallery.com/

arX

iv:1

706.

0578

6v3

[cs

.IR

] 2

3 O

ct 2

017

Page 2: Exploring Content-based Artwork Recommendation with ...processing systems. 1097–1105. [8] Augusto Q Macedo, Leandro B Marinho, and Rodrygo LT Santos. 2015. Context-aware event recommendation

Conference’17, July 2017, Washington, DC, USA Messina et al.

Figure 1: t-SNE map of the DNN image embedding displaying paintings of an anonymized user pro�le (le�, green), and rec-ommendations with three methods: DNN (blue), Metadata (red) and EVF (yellow). Check marks indicate correct predictions.

To help users of the pla�form to explore the vast amount of art-works more e�ciently, they are currently investigating with usthe possibility of top-n content-based recommendation methodswithin the pla�form exploiting features such as artwork metadata,implicit and explicit visual features.

3 DATASETUGallery provided us with an anonymized dataset of 1, 371 users,3, 490 items and 2, 846 purchases (transactions) of paintings, whereall users have made at least one transaction. In average, each userhas bought 2-3 items in the latest years2.

Metadata. Artworks in the UGallery dataset were manuallycurated by experts. In total, there are �ve a�ributes: color (e.g. red,blue), subject (e.g. sports, travel), style (e.g. abstract, surrealism),medium (e.g. oil, acrylic), and mood (e.g. energetic, warm).

Visual Features. For each image representing a painting in thedataset we obtain features from an AlexNet DNN [7], which outputsa vector of 4,096 dimensions. We also obtain a vector of explicitvisual features of a�ractiveness, based on the work of San Pedro etal. [11]: brightness, saturation, sharpness, entropy, RGB-contrast,colorfulness and naturalness.4 EXPERIMENTAL SETUP & RESULTSRecommendation Methods. We compare �ve methods based onthe features used: (1) Metadata: features based on the metadata ofthe items previously bought by the user, (2) DNN : features fromimages using the AlexNet DNN [7], (3) EVF : Explicit visual featuresbased on a�ractiveness of the images [11], (4) Hyb (DNN + EVF):hybrid model using DNN and EVF features, and (5) Hyb (DNN + EVF+ Metadata): hybrid model using DNN, EVF and metadata. For thehybrid recommendations, we combine scores of di�erent sourcesusing the BPR framework [10]. In Figure 1 we see, for instance, auser pro�le at the le� side, besides the image embedding based onfeatures from AlexNet DNN, and then recommendation obtainedby three di�erent methods.

Evaluation. Our protocol is based on the one as introduced byMacedo et al. [8] to evaluate recommender system accuratly in atemporal manner. We a�empt to predict the items purchased inevery transaction, where the training set contains all the artworkspreviously bought by a user just before making the transaction tobe predicted. Users who have purchased exactly one artwork wereremove as their would be no training instance available. Metrics.2Our collaborators at UGallery requested us not to disclose the exact dates when thedata was collected.

Table 1: Results of the simulated recommendation experi-ment.

name ndcg@5 ndcg@10 rec@5 rec@10 prec@5 prec@10Hyb(DNN+EVF+Metadata) .0841 .0990 .1119 .1560 .0279 .0195Hyb(DNN+EVF) .0753 .0934 .0965 .1492 .0235 .0186DNN .0810 .0968 .1052 .1525 .0269 .0195EVF .0370 .0453 .0585 .0826 .0152 .0109Metadata .0312 .0412 .0474 .0773 .0113 .0092

As suggested by Cremonesi et al. [4] for top-n recommendations,we used recall@k and precision@k , as well as nDCG [9].

Results. Table 1 presents the results, which can be summarizedas follows: (1) Visual features outperform metadata features. �isresult is a quite positive �nding as manually cra�ed metadata coststime and money, (2) visual features obtained from the AlexNetDNN perform be�er than those based on explicit visual features.Although this result shows that DNNs do again a remarkable jobin this domain, we are not too happy about it. Features obtainedfrom an DNN such as AlexNet are latent, i.e., we cannot interpretthem directly and we can not use them to explain the recommenda-tions made [12]. Finally, (3) our experiments reveal that the hybridmethod performs even best.5 CONCLUSIONSIn this work we introduce content-based recommendation for physi-cal artworks, comparing manually-curated metadata, AlexNet DNNfeatures, and a�ractiveness-based visual features. Furthermore, weshow that the DNN features outperform the explicit visual featuresand metadata. In practice this has two implications: First, there isno need to exploit metadata as visual features work be�er. Second,it will be di�cult to provide explanations to users as explicit fea-tures work signi�canly worse than latend features obtain via DNNs.It would be interesting though to investigate, whether this gap canbe closed in a real-world experiment. �e current investigations arejust based on simulations and neglect the user factor, though give ahint towards the performance of the models when no explanationsare given.

REFERENCES[1] Xavier Amatriain. 2013. Mining large streams of user data for personalized

recommendations. ACM SIGKDD Explorations Newsle�er 14, 2 (2013), 37–48.[2] LM Aroyo, Y Wang, R Brussee, Peter Gorgels, LW Rutledge, and N Stash. 2007.

Personalized museum experience: �e Rijksmuseum use case. In Proceedings ofMuseums and the Web.

[3] Oscar Celma. 2010. Music recommendation. In Music Recommendation andDiscovery. Springer, 43–85.

Page 3: Exploring Content-based Artwork Recommendation with ...processing systems. 1097–1105. [8] Augusto Q Macedo, Leandro B Marinho, and Rodrygo LT Santos. 2015. Context-aware event recommendation

Exploring Content-based Artwork Recommendation with Metadata and Visual Features Conference’17, July 2017, Washington, DC, USA

[4] Paolo Cremonesi, Yehuda Koren, and Roberto Turrin. 2010. Performance ofRecommender Algorithms on Top-n Recommendation Tasks. In Proceedings ofthe Fourth ACM Conference on Recommender Systems (RecSys ’10). ACM, NewYork, NY, USA, 39–46.

[5] Abigail R. Esman. 2012. �e World’s Strongest Economy? �eGlobal Art Market. h�ps://www.forbes.com/sites/abigailesman/2012/02/29/the-worlds-strongest-economy-the-global-art-market/. (2012). [Online; ac-cessed 21-March-2017].

[6] Ruining He, Chen Fang, Zhaowen Wang, and Julian McAuley. 2016. Vista: AVisually, Socially, and Temporally-aware Model for Artistic Recommendation.In Proceedings of the 10th ACM Conference on Recommender Systems (RecSys ’16).ACM, New York, NY, USA, 309–316.

[7] Alex Krizhevsky, Ilya Sutskever, and Geo�rey E Hinton. 2012. Imagenet classi�ca-tion with deep convolutional neural networks. In Advances in neural informationprocessing systems. 1097–1105.

[8] Augusto Q Macedo, Leandro B Marinho, and Rodrygo LT Santos. 2015. Context-aware event recommendation in event-based social networks. In Proceedings ofthe 9th ACM Conference on Recommender Systems. ACM, 123–130.

[9] Christopher D Manning, Prabhakar Raghavan, Hinrich Schutze, et al. 2008. Intro-duction to information retrieval. Vol. 1. Cambridge university press Cambridge.

[10] Ste�en Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-�ieme. 2009. BPR: Bayesian personalized ranking from implicit feedback. InProceedings of the twenty-��h conference on uncertainty in arti�cial intelligence.AUAI Press, 452–461.

[11] Jose San Pedro and Stefan Siersdorfer. 2009. Ranking and Classifying A�ractive-ness of Photos in Folksonomies. In Proceedings of the 18th International Conferenceon World Wide Web (WWW ’09). ACM, New York, NY, USA, 771–780.

[12] Katrien Verbert, Denis Parra, Peter Brusilovsky, and Erik Duval. 2013. Visualizingrecommendations to support exploration, transparency and controllability. InProceedings of the 2013 international conference on Intelligent user interfaces. ACM,351–362.