Chris Holder
Convolutional networks for appearance based recommendation and visualisation of mascara products
Holder, Chris; Ricketts, Stephen; Obara, Boguslaw
Authors
Stephen Ricketts
Boguslaw Obara
Abstract
In this work, we explore the problems of recommending and visualising makeup products based on images of customers. Focusing on mascara, we propose a two-stage approach that first recommends products to a new customer based on the preferences of other customers with similar visual appearance and then visualises how the recommended products might look on the customer. For the initial product recommendation, we train a Siamese convolutional neural network, using our own dataset of cropped eye regions from images of 91 female subjects, such that it learns to output feature vectors that place images of the same subject close together in high-dimensional space. We evaluate the trained network based on its ability to correctly identify existing subjects from unseen images, and then assess its capability to identify visually similar subjects when an image of a new subject is used as input. For product visualisation, we train per-product generative adversarial networks to map the appearance of a specific product onto an image of a customer with no makeup. We train models to generate images of two mascara formulations and assess their capability to generate realistic mascara lashes while changing as little as possible within non-lash image regions and simulating the different effects of the two products used.
Citation
Holder, C., Ricketts, S., & Obara, B. (2020). Convolutional networks for appearance based recommendation and visualisation of mascara products. Machine Vision and Applications, 31, Article 5. https://doi.org/10.1007/s00138-019-01053-5
Journal Article Type | Article |
---|---|
Acceptance Date | Dec 3, 2019 |
Online Publication Date | Jan 21, 2020 |
Publication Date | 2020 |
Deposit Date | Dec 3, 2019 |
Publicly Available Date | Jan 22, 2020 |
Journal | Machine Vision and Applications |
Print ISSN | 0932-8092 |
Electronic ISSN | 1432-1769 |
Publisher | Springer |
Peer Reviewed | Peer Reviewed |
Volume | 31 |
Article Number | 5 |
DOI | https://doi.org/10.1007/s00138-019-01053-5 |
Public URL | https://durham-repository.worktribe.com/output/1282357 |
Files
Published Journal Article
(3.3 Mb)
PDF
Publisher Licence URL
http://creativecommons.org/licenses/by/4.0/
Copyright Statement
This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
You might also like
Exploring the semantic content of unsupervised graph embeddings: an empirical study
(2019)
Journal Article
The multiscale bowler-hat transform for blood vessel enhancement in retinal images
(2018)
Journal Article
Multi-scale Segmentation and Surface Fitting for Measuring 3D Macular Holes
(2017)
Journal Article
Downloadable Citations
About Durham Research Online (DRO)
Administrator e-mail: dro.admin@durham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search