Face 2 Emoji
Face 2 Emoji
Author Keywords
Permission to make digital or hard copies of part or all of this work for personal or Face2Emoji; emoji; crowdsourcing; emotion recognition;
classroom use is granted without fee provided that copies are not made or distributed
for profit or commercial advantage and that copies bear this notice and the full citation facial expression; input; keyboard; text entry
on the first page. Copyrights for third-party components of this work must be honored.
For all other uses, contact the owner/author(s). Copyright is held by the
author/owner(s).
ACM Classification Keywords
CHI’17 Extended Abstracts, May 06–11, 2017, Denver, CO, USA. H.5.m [Information interfaces and presentation (e.g., HCI)]:
ACM 978-1-4503-4656-6/17/05.
http://dx.doi.org/10.1145/3027063.3053086
Miscellaneous
Introduction Motivation & Research Questions
Nonverbal behavior conveys affective and emotional infor- Face2Emoji is motivated by two findings from the literature:
mation, to communicate ideas, manage interactions, and that a primary function of emojis is to express emotion, and
disambiguate meaning to improve the efficiency of con- that most emojis used are face emojis. Cramer et al. [3]
versations [14, 25]. One way to indicate nonverbal cues is found that 60% (139/228) of their analyzed message by
by sending emoji, which are graphic icons (e.g., , , ) US participants were emoji used for expressing emotion.
managed by the Unicode Consortium1 that are identified by In an Instagram emoji study4 , faces accounted for 6 of the
unicode characters and rendered according to a platform’s top 10 emojis used, providing further evidence that peo-
font package. ple frequently use emoji to express emotion. Furthermore,
according to a 2015 SwiftKey report5 , faces accounted for
Emojis enable people to express themselves richly, and close to 60 percent of emoji use in their analysis of billions
while shown as screen graphics, they can be manipulated of messages. Finally, in a qualitative study from Lee et al.
as text structures. Besides Pohl et al.’s EmojiZoom [22] [17] on emoticon sticker usage, they found that these stick-
who propose a zooming-based interface, entering emoji ers were used mainly for expressing emotions.
on smartphone keyboards currently requires users to make
a selection from large lists (one list per category of emoji) The study of nonverbal communication via emotions origi-
(e.g., Apple© iOS 10 emoji keyboard2 in Fig. 1). This makes nated with Darwin’s claim that emotion expressions evolved
emoji entry “a linear search task" [22], and given the grow- in humans from pre-human nonverbal displays [4]. Further-
Figure 1: Apple© iOS 10 emoji ing number of emojis, we assume can incur user frustration. more, according to Ekman [6, 7], there are six basic emo-
keyboard within iMessage.
While no prior work explicitly addresses this, efforts such as tions which have acquired a special status among the sci-
Emojipedia3 highlight the need for better emoji search. entific community: Anger, Disgust, Fear, Happiness, Sad-
ness, and Surprise. Here, we draw on these six basic emo-
To address this, we propose Face2Emoji, a system and tions, and additionally include the Neutral facial expression.
method to use users’ facial emotional expressions as sys- By using computer vision and machine learning techniques
tem input to filter emojis by emotional category. Despite that for analyzing and recognizing emotional expressions, the
emojis can represent actions, objects, nature, and other user’s face can be used as a natural interaction filter6 . To
symbols, the most commonly used emojis are faces which test the validity of our proposed method, we used crowd-
express emotion [3, 17, 24]. Moreover, previous work has sourcing to firstly identify whether a natural mapping be-
shown that emojis can be ranked by sentiment (cf., Emoji
Sentiment Ranking by Novak et al. [15]), textual notifica- 4
https://www.tumblr.com/dashboard/blog/instagram-
tions containing emojis exhibit differences in 3-valued sen- engineering/117889701472 ; last retrieved: 14-02-2017
timent across platforms [23], and for faces, emojis can be 5
https://blog.swiftkey.com/americans-love-skulls-brazilians-love-cats-
ranked by valence and arousal [24]. swiftkey-emoji-meanings-report/ ; last retrieved: 14-02-2017
6
A filter according to Wikipedia (https://en.wikipedia.org/wiki/Filter_(higher-
1
http://unicode.org/emoji/ ; last retrieved: 14-02-2017 order_function)) is defined as “a higher-order function that processes a
2
Source: https://support.apple.com/en-us/HT202332 ; last retrieved: data structure (usually a list) in some order to produce a new data struc-
14-02-2017 ture containing exactly those elements of the original data structure for
3
http://emojipedia.org/ ; last retrieved: 14-02-2017 which a given predicate returns the boolean value true."
tween emojis and the seven facial expressions exists, and if municate [18]. In a study by Barbieri et al. [2], they found
so, what this mapping distribution looks like. that the overall semantics of the subset of the emojis they
studied is preserved across US English, UK English, Span-
We address the following questions: Do the most frequently ish, and Italian. As validation of the usefulness of mapping
used emojis naturally map to the six basic (+ Neutral) facial emojis to emotions, preliminary investigations reported by
emotional expressions? Can we achieve reasonable facial Jaeger et al. [13] suggest that emoji may have potential as
emotional expression recognition for these emotions using a method for direct measurement of emotional associations
deep convolutional neural networks? The rest of the paper to foods and beverages.
will address related work on natural, multimodal user inter-
faces and emoji communication, provide our crowdsourcing Emoji (Mis-)interpretation
approach and results, our early emotion recognition ex- Recently, Miller et al. [20] demonstrated how same emoji
periments using deep convolutional neural networks, and look differently across devices (iPhone, Android, Samsung)
sketch our future research steps and open questions. and is therefore differently interpreted across users. Even
when participants were exposed to the same emoji render-
Related Work ing, they disagreed on whether the sentiment was positive,
Multimodal User Interfaces and Emoji Entry neutral, or negative around 25% of the time. In a related
Related to our approach, Filho et al. [8] augmented text preliminary study, Tigwell et al. [24] found clear differences
chatting in mobile phones by adding automatically detected in emoji valence and arousal ratings between platform pairs
facial expression reactions using computer vision tech- due to differences in their design, as well as variations in
niques, resulting in an emotion enhanced mobile chat. For ratings for emoji within a platform. In the context of our
using the user’s face as input, Anand et al. [1] explored a work, this highlights the need to account for multiple in-
Face2Emoji: Mapping facial emotional expressions to Emoji use-case of an eBook reader application wherein the user terpretations, where an emoji (as we show later) can be
Which emotional expression is this emoji most associated with?
Progress: 1/202 performs certain facial expressions naturally to control the classified as belonging to one or more emotion categories.
device. With respect to emoji entry, Pohl et al. [22] pro-
posed a new zooming keyboard for emoji entry, EmojiZoom, Crowdsourcing Emoji to Emotion Mappings
where users can see all emoji at once. Their technique, Approach
Afraid Angry Disgusted Neutral Happy Sad Surprised
which was tested in a usability study against the Google To validate whether emojis, irrespective of function, can
Skip: emoji not displayed correctly
keyboard, showed 18% faster emoji entry. be categorized into one of the six basic emotional expres-
Instructions
Click / Tap on the facial expression that you think best matches the emoji shown. If the
emoji is not displayed correctly, then choose 'Skip'.
sions (+ Neutral), and what such a mapping looks like, we
Emoji and Emotion Communication adopted a crowdsourcing approach. Since currently as of
The compactness of emojis reduces the effort of input to Unicode 9.0, there are 1,394 emojis (not including modified
PDF generated automatically by the HTML to PDF API of PDFmyURL
Figure 2: Snapshot of the express not only emotions, but also serves to adjust mes- emojis, or sequences)7 , we decided to test only a subset.
Face2Emoji crowdsourcing website sage tone, increase message engagement, manage con- We selected emojis with greater than 100 occurrences from
(showing female faces here). versations and maintain social relationships [3]. Moreover, the Emoji Sentiment Ranking V1 [15] dataset, which re-
emojis do not have language barriers, making it possible for sulted in a total of 202 emojis.
users across countries and cultural backgrounds to com-
7
http://emojipedia.org/stats/ ; last retrieved: 14-02-2017
We built a website to crowdsource emoji to emotion labels into emotions (κ=0.55, CI: [0.46,0.65]), where we agreed
(shown in Fig. 2). On the website, an emoji would be shown on 71.3% (144/202) of emojis. These joint labels were
and a visitor has to choose one of seven emotion faces8 : then compared with the top ranked (majority voted) emo-
Afraid, Angry, Disgusted, Neutral, Happy, Sad, Surprised. jis from the crowd, which gave an almost perfect agreement
Additionally, a ‘Skip’ option was provided in case the emoji (κ=0.85, CI: [0.77,0.93]).
was not displayed correctly. We tracked emojis and emo-
tion labels using cookie session IDs, where the emoji index Classification Results
and associated unicode were used for all subsequent anal- The distribution of the top most frequent (by majority vote)
Operating System Labels ysis. We additionally tracked a visitors’ Operating System, emotion labels, as well as the next top labels, across the
Win32 7113
however not the browser type (which can be a limitation). 202 tested emojis are shown in Fig. 3. Interesting to ob-
MacIntel 3033
iPhone 2347 IP addresses were not tracked to avoid data privacy issues. serve here that for the majority of labels, none of the emojis
Android 1269
Furthermore, we chose to render the unicode and not cre- tested were skipped due to unicode rendering. From our
Linux 517
iPad 449 ate images from them, in order to ensure users across plat- labeled data, it became clear that an emoji can be classi-
Win64 427
forms can provide emotion labels, irrespective of rendering. fied under two emojis (following a bimodal or at times mul-
The website was distributed via online forums (e.g., Red- timodal distribution). For example, was nearly equally
Table 1: OS’s used to access the
dit) and the authors’ social networks. Our datasets (raw and labeled as Happy (N=32) (since a trophy, a sign of achieve-
Face2Emoji website across all
ranked) are publicly available9 for research purposes here: ment can evoke happiness) and Neutral (N=34), since
visitors (N=15,155)
https://github.com/abdoelali/f2e_dataset it is an object with no direct mapping to a facial expres-
100 sion. Therefore, to account for this variability, we classified
80 Descriptive Statistics whether an emoji belongs to an emotion label using our
We collected a total of 15,155 labels, across 308 unique Emotion Class (EC) function:
Frequency
60
40
website visitors. Each emoji received an average of 75.0
labels (M d=74.5, s=5.3). From the total set, 1,621 (10.7 %)
20
were ‘skipped’ (or labeled as NA’s), where 10% of respon-
(
xij 1 if EC > 0.5
0
afraid angry disgusted happy na neutral sad surprised dents who labeled NA made up 73.3% (1188/1621) of all EC = = (1)
Emotions
NAs in our dataset. The distribution of operating systems
max (xi ) 0 if EC ≤ 0.5
top next top
used to access the website are shown in Table 1.
Figure 3: Distribution of top and
Annotation Quality where: xi ∈ [1, 202], xj ∈ [1, 8]. We chose a cutoff thresh-
next top crowdsourced majority
As a test for annotation quality, we independently (N=2) old of 0.5, where an emoji is classified as belonging to an
voting of 202 emojis across
emotion categories, including NAs. rated each emoji by classifying into one of the emotion cat- emotion class if EC > 0.5. The result of applying our EC
egories, and computed unweighted Cohen’s Kappa. Our function to our data is shown in Table 2, where the emojis
ratings reached moderate agreement on classifying emojis per emotion category are sorted by label count in ascend-
ing order.
8
Female or male faces randomly chosen on page refresh.
9
Our datasets contain no sensitive information and therefore comply
with user privacy.
6000
5500
5000
4500
4000
Frequency
Emotions
training validation
0.6 certain emotions are not correctly classified, (b) we can en- seen that most of the emotions are Happy, Neutral, Sad,
sure user privacy by running all predictions directly on the and Surprised. After experimenting with different architec-
0.4
device, and (c) it is free. ture and hyperparameters, our final network architecture
0 25 50 75 100 is shown in Table 3, where training was done with a batch
Dataset & Architecture
Epochs size of 32, using stochastic gradient descent with hyper-
acc val_acc We used the FER-2013 facial expression dataset [9] for
parameters (momentum=0.9, learning rate= 0.001, weight
training and validation, which comprises 32,298 grayscale
Figure 5: Accuracy and validation 10 11
https://www.microsoft.com/cognitive-services/en-us/emotion-api ; last http://tflearn.org/ ; last retrieved: 14-02-2017
of our final network model across 12
retrieved: 14-02-2017 https://www.tensorflow.org/ ; last retrieved: 14-02-2017
100 epochs.
decay=0.0005) where loss was computed using categori- tend to test our Face2Emoji app with users for only those
neutral 0.03 0.01 0.02 0.04 0.10 0.02 0.78 cal cross-entropy, and run on an NVIDIA GeForce GTX 970 three emotions: Happy, Surprised, and Neutral.
surprised 0.01 0.01 0.08 0.03 0.03 0.81 0.04
GPU for 100 epochs.
Next Steps & Research Agenda
sad 0.03 0.04 0.05 0.06 0.43 0.04 0.34 Early Experiments & Results Our next steps are to experiment further with deep learning
Accuracy and validation accuracy plots across 100 epochs approaches for emotion recognition (e.g., using transfer-
Real Emotion