Intelligent Clothing Interaction Design and Evaluation System Based on DCGAN
Intelligent Clothing Interaction Design and Evaluation System Based on DCGAN
Conference Series
1
Clothing Institute, Shanghai University of Engineering Science, Shanghai, 201620,
China
Abstract. The deep convolutional generative adversarial network (DCGAN) model can be
applied to personalized fashion design. Determining the best design from fashion sketches that
is of interest to customers is an important research field. In this study, we propose a rapid
evaluation method based on Kansei engineering to evaluate personalized fashion images
generated by DCGAN. The degree of customer interest in the generated fashion images is
determined by tracking the eye movements and facial expressions of customers. Images with
high number of fixations are added to the training data for iterative training; this aims to
improving customer satisfaction. The method was verified in an interaction design of
personalized black dresses. After several iterations, the customers’ satisfaction rating of the
generated images by DCGAN improved. This method can be applied to other styles of fashion
design.
1. Introduction
At present, research on the image generation model is closely related to the fashion image generation
technology. The image generation technology can be generally divided into two categories: generative
parametric approaches and adversarial approaches[1]. Various studies have shown that adversarial
approaches have clear advantages in image generation, especially the generative adversarial network
(GAN). The generation of deep convolutional GAN (DCGAN) algorithm[2] is a milestone in the
development of the GAN model, which determines a stable training structure to facilitate engineering
implementation.
In recent years, research on GANs has increased, and related research has made some progress. For
example, Phillip et al.[3] used the conditional GAN algorithm as an image conversion solution that
effectively implements the conversion of linear drawings to rendered images. Jun-Yan Zhu et al.[4] also
carried out a research on image style migration through the CycleGAN algorithm. Yoo D et al.[1] set a
converter in the generative model to generate an image of a person wearing clothes. Makkapati V et
al.[5] put forward related theories on the symmetry of generated fashion images. Although the
aforementioned research has made contributions to the application of deep learning in fashion design,
these studies have not considered users’ personalized needs. Therefore, generated fashion images
should be evaluated and selected from the perspective of customers, and explorations should be made
in terms of improving customer satisfaction.
Currently, in addition to traditional methods, Kansei engineering has also been widely used[6]. For
example, eye tracking[7] can reflect customers’ attraction on a certain product. The longer the fixation
Content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution
of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.
Published under licence by IOP Publishing Ltd 1
SIFTIC 2020 IOP Publishing
Journal of Physics: Conference Series 1790 (2021) 012025 doi:10.1088/1742-6596/1790/1/012025
time, the more interested the customer is in the design[8]. In addition, facial expressions are the most
important, natural, and direct channels for people to express emotions[9] .
This study creatively applies deep learning algorithms to automatically generate fashion designs
and Kansei engineering for the evaluation of costumers’ preferences, achieving a fusion of Kansei
engineering and fashion design. This paper is organized as follows: In Section 2, the framework of the
system proposed in this study is presented. In Section 3, the experimental procedure and results are
presented to show the effectiveness of our method. In Sections 4 and 5, a discussion and future
outlook are given, respectively.
The dashed box represents the DCGAN network that automatically generates design modules.
Outside the dashed box, the customer quickly selects the evaluation module. The specific steps are
described as follows:
2.2. Selection
After the training of the DCGAN network was completed, we used the DCGAN module to generate
the fashion design images. When the fashion designs were automatically and continuously played in
the form of a slideshow, the user showed her response to the related designs and screened out the
clothing designs that are positively attracted.
2.3. Replacement
The fashion design images that the customer selected through the interactive evaluation module were
used as new training data to replace part of the original training data after passing through the data
augmentation module, and the design images not selected were discarded.
3. Experiment
2
SIFTIC 2020 IOP Publishing
Journal of Physics: Conference Series 1790 (2021) 012025 doi:10.1088/1742-6596/1790/1/012025
This study used little black dress images as a training data set. The specific processes of the
experiment are presented as follows:
3.1. Procedure
3.1.1. Construction of the DCGAN model to generate little black dresses DCGAN already has a ready-
made code in Tensorflow:carpedm20/DCGAN-tensorflow; we used this code directly. We found a few
lines of code to read data, as shown in Figure 2. Then, we created a new anime folder in the data folder,
placed little black dress images into this folder, and specified the dataset anime at runtime. The
training environment is win10+cuda9.2+Tensorflow.
3.1.2. Generate personalized little black dress designs We selected two female college students, aged
23 and 25 years old, as the experimental subjects, and named them S1 and S2 hereinafter, respectively.
For each customer, we obtained 20 little black dress images from their recent online browsing records
and added the images to the training data. Then, we let them choose their top 20 favorite little black
dresses to be added into the training data. The last 60 little black dress images, which were randomly
selected from the database, were added to the training data. The 100 little black dresses were
augmented to 1,000 images and normalized to a 28 × 28 × 1 size. Figure 3 presents the selection of the
initialization training data set.
In this manner, each of the two customer has their own training data set. Accordingly, we named
two groups of modules, i.e., DCGAN-S1 and DCGAN-S2, as shown in Figure 4. The training
environment is win10 + cuda9.2 + Tensorflow.
3
SIFTIC 2020 IOP Publishing
Journal of Physics: Conference Series 1790 (2021) 012025 doi:10.1088/1742-6596/1790/1/012025
After the training, we randomly selected 25 images through the DCGAN network, removed five
pictures with inferior pictures, and made the remaining 20 pictures into continuous slides. The Tobii
eye tracker was used to collect the eye tracking data of customer S1, whereas the FaceReader software
(launched by VicarVision and Noldus Information Technology) was used to measure and analyze the
facial expressions of customer S2. Before the experiment, the customers were informed that for the
research study, expression video materials need to be collected; they were ensured of data
confidentiality. We recorded the number of designs that the customers tended to positively select (i.e.,
customers’ favorite designs) in each round, as shown in Figures 5 and 6.The selected design images
replaced some of the training images.
3.2. Results
Table 1 presents the number of images selected by each experimental subject through the interactive
evaluation module in the first four rounds. The table shows that after three rounds of iterative training
of the DCGAN, the proportion of selected generated images increased.
Table 1. Selected images and percentage of selection (PoS) after four rounds of
iterative training of the two customers.
S1 Pos S2 Pos
5. Future outlook
This research creatively applies deep learning algorithms to automatically generate clothing designs,
achieving a fusion between Kansei engineering and fashion design. In the future, more experiments
should be performed to verify the effectiveness of this system. In addition, the following
improvements will be made:the specific effects of eye tracking and facial expression analysis on
4
SIFTIC 2020 IOP Publishing
Journal of Physics: Conference Series 1790 (2021) 012025 doi:10.1088/1742-6596/1790/1/012025
customers’ fashion design preferences should further examined;training data should be updated in a
timely manner;customers’ selected design images should be used to make real clothing; fashion
designers will benefit from this type of system.
Acknowledgments
2
This research was funded by the Program for Professor of Special Appointment (Eastern Scholar) at
Shanghai Institutions of Higher Learning (Grant No. TP2017074).
References
[1] Yoo D, Kim N, Park S, et al. Pixel-level domain transfer. European Conf. on Computer
Vision. 2016 (Springer, Cham)
[2] Radford A, Metz L and Chintala S. Unsupervised Representation Learning with Deep
Convolutional Generative Adversarial Networks 2015 J. Computer Science
[3] Isola P, Zhu J Y, Zhou T, et al. Image-to-Image Translation with Conditional Adversarial
Networks IEEE Conf. on computer vision and pattern recognition. 2017
[4] Zhu J Y, Park T, Isola P, et al. Unpaired Image-to-Image Translation using Cycle-Consistent
Adversarial Networks IEEE Conf. on computer vision and pattern recognition. 2017
[5] Makkapati V and Patro A. Enhancing Symmetry in GAN Generated Fashion Images
International Conf. on Innovative Techniques and Applications of Artificial Intelligence.
2017 (Springer, Cham)
[6] Li F, Liu X G and Gu W. Evaluation process of costume design based on evaluation subject
perspective 2017 J. Journal of Textile
[7] Li Zh, Gou B C, Chu J J, et al. Way of getting user requirements based on eye tracking
technology. 2015 J. Computer Engineering and Applications 51 9 233–7
[8] Jiang Q, Peng L, Gao Y K, Yang X Y and Huang W Extraction and research oncultural
characteristics of Longhorn Miao costume and adornment 2019 J. Modern Electronics
Technique 42 14 135–9
[9] Keltner D and Ekman P. 2010 Facial Expression of Emotion Handbook of Emotions 3rd.
(New York: The Guilford Press) pp 173–83