Preprint
Article

Are Men Better at Recognizing Emojis than Facial Expressions?

This version is not peer-reviewed.

Submitted:

22 January 2023

Posted:

23 January 2023

You are already at the latest version

A peer-reviewed article of this preprint also exists.

Abstract
Emojis are colorful ideograms resembling stylized faces commonly used for expressing emotions in instant messaging, in social network sites and in email communication. Notwithstanding their increasing and pervasive use in electronic communication, they are not much investigated in terms of their psychological properties and communicative efficacy. Here we presented 112 different human facial expressions and emojis (expressing neutrality, joy, surprise, sadness, anger, fear and disgust) to a group of 96 female and male university students engaged in the recognition of their emotional meaning. Both Analysis of Variance and Wilcoxon tests showed that men were significantly better than women at recognizing emojis (especially negative ones) while women were better than men at recognizing human facial expressions. Quite interestingly, men were better at recognizing emojis than human facial expressions per se. These findings are in line with more recent evidences suggesting how men may be more competent and inclined to use emojis to express their emotions in messaging (especially sarcasm, tease and love) than previously thought. Finally, the data indicate how emojis are less ambiguous than facial expressions (except for neutral and surprise emotions), possibly because of the limited number of fine-grained details, and the lack of morphological features conveying facial identity.
Keywords: 
Subject: 
Social Sciences  -   Cognitive Science

1. Introduction

Emojis are colorful ideograms, slightly larger than letters, which can represent facial expressions (or other entities such as hand signs, food, animals, etc…) and are typically used in computer- or mobile-mediated communication, particularly in instant messaging systems and social networks. Emojis may be accompanied by written text and are intended to replace nonverbal communication as much as possible [1]. They are effectively used in textual communication to compensate for the lack of nonverbal signs and to enrich communication emotionally, promoting the expression of emotions. Some studies, along with the extreme pervasiveness of this modern communication system in messaging, have demonstrated their effectiveness. For example, Aluja and coauthors [2] measured the acoustic startle reflex modulation during observation of emoji emotional faces and found higher acoustically evoked startle responses when viewing unpleasant emoji and lower responses for pleasant ones, similarly to what obtained with real human faces as emotional stimuli. Again, Cherbonnier and Michinov [3] compared emotion recognition mediated by different types of facial stimuli: pictures of real faces, emojis, and face drawings and found that, contrary to what one might suppose, participants were more accurate in detecting emotions from emojis than real faces and emoticons. Liao et al. [4] (2021), comparing ERPs generated by viewing painful and neutral faces and emojis, found that P300 was of greater amplitude in response to faces than emojis, but both types of stimuli elicited late positive potentials (LPP) of greater amplitude in response to negative than positive emotions. The authors concluded that emojis and faces were processed similarly, especially at the later cognitive stages.
Emojis and emoticons have become a worldwide means for expressing emotions in computer-mediated communication, but neuroscientific knowledge regarding how the human brain recognizes them is still scarce. Some studies were performed to compare face and emoticon processing. The results seem to suggest a partial independence in the subserving neural circuits. Yuasa and coworkers [5] compared cerebral activations during observation of Japanese faces and Japanese emoticons (e.g., (^_^) combinations of vertically organized letters and symbols) through fMRI and found that the face fusiform area (FFA) was active only during processing of human faces (as expected, e.g., [6]), and not of emoticons. However, the right inferior frontal gyrus, involved in the assessment of stimulus emotional valence, was active during processing of both types of stimuli. This was interpreted as an indication that, even if emoticons were not perceived as faces, they activated higher order social areas devoted to the interpretation of face emotional meaning. Similarly, using emoji stimuli, Chatzichristos and coworkers [7] failed to find an FFA activation during the perception and matching of positive and negative emojis with words triggering autobiographical memories. Instead, it was found the activation of a neural circuit implicated in emotional and social processing, including the inferior frontal gyrus, the amygdala, and the right temporal pole. However, other neuroimaging studies have found the activation of the FFA and the Occipital face area (OFA, devoted to face part analysis) during observation of emoticons, which would explain how they would be emotionally recognized and interpreted as facial stimuli [8-10].
The purpose of this study was manifold. First, it was assessed the communicative effectiveness of sets of emoji and face stimuli through preliminary validation. Then, it was investigated to which extent observers grasped the emotional meaning of real facial expressions and emojis, as a function of stimulus type, valence and complexity. Another important aim of the study was to test the existence of sex differences in the ability to recognize the affective content of faces vs. emojis. Numerous studies have provided evidence that women do better than men in task involving deciphering emotions through facial expressions and nonverbal communication cues [11-14]. Women would also be more likely than men to express their emotional experiences to others [15]. This sex difference would be more pronounced with subtle emotions (such as sadness or fear), also would also reflect a greater interest of females in social information [12] and a more empathic attitude toward suffering of others [16].
On the other hand, several studies pointed to gender differences in emoji use. Women, in fact, according to some authors, would use emojis more frequently than men [17]. In line with this evidence, Jones and coworkers [18] assessed the familiarity and the perceived valence for 70 facial emojis in a large sample of United States college students and found an higher emoji usage and familiarity ratings for women than for men. Since emojis were introduced in the on-line text communication to express sentiments and describe more effectively emotional information, it could be presumed that women would manifest a more intense use of emojis and emoticons than men, similarly to what is observed in real social interactions [19-20]. However, this does not seem to be the case. Chen and coworkers [21] documented that 7.02 % of male users used at least one emoji in their typical messages while 7.96% of female were likely to use one or more emojis. Herring & Dainas [22] reported that females and males social media users interpreted and used similarly emojis in Facebook messages. However, they were surprised to find that male users were significantly more likely to use heart-related emojis than females (female: 17.62%, male: 19.41%, p-value = 0.01). This is contrary to psychological literature where males are reported to be less willing to express love in real life [23]. According to the authors, such a finding would imply that, although men reserve to express their love in real life, they are more willing to express love through emojis in textual communication. After all, it is possible that social behaviors change as technological contexts change.
To investigate these issues we compared the degree of recognizability of human facial expressions relative to eight different human identities (half male, half female) with that of emojis printed in eight different graphic styles, in a large sample of students. Seven different neutral, positive and negative facial expressions were considered. On the basis of the previous literature it was expected that women outperformed men in the recognition of empathy related, subtle, human facial expressions (e.g., fear) [24]. In addition, it was generally hypothesized that emoji and face recognizability scores would be roughly comparable, with some occasional advantage for emojis, depending on the specific emotion [25-26]. The artificiality of this symbolic system would in fact be counterbalanced by emojis’ lack of ambiguity, limited number of fine-grained features, and lack of internal noise related to the variety of human identities.

2. Materials and Methods

2.1. Participants

96 healthy students of local University (48 males and 48 females) aged 18 to 35 years (mean age = 23.89) participated in the study. They were randomly assigned to the facial expression study (48 subjects: 24 males, aged: 23.37; 24 females: 23.37 years) or the emoji recognition study (48 subjects: 24 males aged 23.62; 24 females: 24.91). The current sample size was tested for power analysis by using the program G*Power3.1 for comparing two independent groups with alpha level = 0.05. They were all right-handed, as assessed through the administration of the Edinburgh Handedness Inventory. They all declared to have never suffered from psychiatric or neurological deficits and to have good or corrected-to-good vision. Before taking part in the experiment, participants signed written informed consent forms. The study was carried out in accordance with the relevant guidelines and regulations and was approved by the ethics committee of University of Milano-Bicocca (CRIP, protocol number RM-2021-401). It was conducted online and programmed via Google forms https://www.google.com/forms.

2.2. Stimuli

56 different human facial expressions and 56 different emoji pictures (i.e. 112 different visual stimuli), depicting 8 variants of 7 universal facial expressions were used in this study. Eight different identities (4 female, 4 male) were used for human faces, while eight different emoji style were used for emoji testing. Stimuli were created and validated as described below.
Emoji. Emojis pictures were drawn from free web platforms (Apple, Google, Microsoft, WhatsApp, Twitter, Facebook and Joypixels and www.icons8.it) and represented the six basic
Ekman’s emotions [27]: joy, sadness, surprise, fear, anger, and disgust plus neutrality, in eight different styles. They were matched for size (4 cm of diameter) and average luminance. The original stimuli were slightly modified to adjust their average luminance, color saturation and size. Figure 1 illustrates some examples of stimuli. Stimuli were preliminarily validated in a behavioral study by means of a test administered via Google Form to select the most easily recognizable emojis among a much larger set of 168 emojis comprising 24 different emoji styles for 7 facial expressions.
Validation was performed on group of 48 students aging on average 23 years. All participants had normal vision, no neurological or psychiatric deficits and possessed diploma, BA or Master degrees. Participants were shown, randomly mixed and once at a time, the various emojis and were required to rapidly observe the picture and decide which, of the seven emotion words typed below, was more appropriate to describe the viewed facial expression, by clicking a check mark within a few seconds. Pictures were displayed at the center of the screen and the experimental session lasted 10 minutes. The eight sets (featuring 56 emojis) associated with the highest accuracy in performance were selected as experimental stimuli. Average hit rate for the final set was 79.40%. In more details, accuracy was 94.53% (SD= 0.51) for joy, 76.04% (SD= 1.85) for surprise, 84.12% (SD= 1.76) for sadness, 96.09% (SD= 2.1) for anger, 57.29% (5.04) for fear, 91.15 % (SD= 0.99) for disgust and 56.51% (SD= 0.99) for neutrality.
Faces. 10 student actors of Caucasian ethnicity were recruited (5 females and 5 males) aging 23 years on average (SD= 1.333) for photos taking. High-resolution pictures of their faces were self- taken with a cell phone at about 40 cm of distance in light controlled conditions, while standing up against a white wall. Actors were required to avoid wearing earrings, glasses, make up, hairpins, pliers, any type of hair embellishments, mustaches, beard. They were also instructed to wear a black t-shirt and gather the hair behind the head. For each of the seven emotions, actors were instructed to imagine a vivid emotional state, while concentrating on a specific autobiographic scenario through the Stanislavsky method, and express it spontaneously. For ‘surprise’ emotion, they were instructed to think of a positive surprise (see Figure 1 for some examples). Each of the 10 actors provided written consent and filled in the privacy release form.
Stimulus set was validated on a group of 50 students (25 females, 24 males and 1 gender fluid) aging on average 23.7 years (min= 17, max= 34 years). All participants had normal vision, no neurological or psychiatric deficits and possessed diploma, BA or Master degrees. Participants were shown, randomly mixed and once at a time, the 56 pictures relative to the seven facial expressions acted by the eight female and male actors. Subjects were required to rapidly observe the picture and decide which one of the seven emotion words typed below was more appropriate to describe the viewed facial expression, by clicking a check mark within a few seconds. Pictures were displayed at the center of the screen and the experimental session lasted 10 minutes. Overall performance for correctly identifying facial emotions was remarkably high = 87.35% (with a chance rate of 16.7%). No participant performed below an overall rate of 75.0%. In more details, accuracy was 98.47% for joy, 86.73% for surprise, 80.1% for sadness, 89.29% for anger, 72.70% for fear, 85.97% for disgust and 98.21% for neutrality. These recognition rates (in line with the data reported by Carbon [28]) outperform the accuracy of recognizing facial expressions reported by other studies in the literature (e.g., 57.85% for anger and disgust in Aviezer et al. [29] and 57.85 for negative emotions in Derntl et al. [30] thus supporting the qualitative validity of the stimuli. Stimulus set was also evaluated for facial attractiveness by a further group of 12 students (7 females and 5 males) aged between 18 to 25 years. Judges were requested to evaluate the attractiveness of neutral expressions of all identities, by using a 3-point Likert scale, where 1 stood for “not attractive”, 2 for “average” and 3 for “attractive”. The results showed a perfect balance across the two sexes and indicated an “average” degree of attractiveness for the facial stimuli (Females= 1.83; SD= 0.78; Males= 1.82; SD= 0.76). This characteristic of stimuli promotes the generalizability of results to the normally looking population. Face stimuli had the same size (3.37 x 5 cm; 199 x 295 pixels; 3° 22’ x 5°).

2.3. Procedure

The emotion-recognition task consisted in 112 experimental trials, in which participants were first shown a portrait photograph of an adult face (or a facial emoji, according to the experiment) to be inspected for about 2 seconds. The images were equiluminant as assessed by subjecting their luminance values to an analysis of variance (F= 0.099, p= 0.992). Photos of faces and emoji were in color, and were displayed at the center of the screen, on a white background. Immediately below the face or the emoji, there was a list of words (neutrality, happiness, surprise, fear, anger, sadness, disgust), from which they had to select the emotion that they deemed the most appropriate to describe the meaning of the expression. In addition, participants judged how clearly they considered the expression recognizable on a 3-point Likert scale (ranging from ‘1 = not much’ to ‘3 = very much’). The emotion was scored 0 if a different incorrect expression was selected. 5 seconds were allowed for perceiving and responding to the two queries. Participants were instructed to observe one facial stimulus at a time and to respond within 5 seconds, not missing any answer. Only one choice per face/emoji was allowed. The task lasted about 15 minutes.

2.4. Data Analysis

The individual scores obtained from each individual, for each of the 7 facial expressions and stimulation condition, underwent a 3-way repeated-measure ANOVA whose factors of variability were: 1 between-group factor named “sex” (with 2 levels, female and male), 1 between-group factor named “stimulus type” (with 2 levels, emoji and face), and “emotion” between-group factor (with 7 levels, happiness, neutrality, surprise, anger, sadness, fear, disgust). Multiple post-hoc comparisons were performed using Tukey's test. Greenhouse-Geisser correction was applied in case of epsilon <1 and epsilon corrected p value were computed. A non-parametric analysis
(Wilcoxon test) was also applied to all the raw data to measure the distribution of individual scores in relation to the sex of viewer and type of stimulus (regardless of facial expression).

3. Results

The ANOVA performed on recognition rates of emojis vs. faces yielded the significance of Emotion (F 6,552 = 31.575, p˂0.00001). Post-hoc comparisons showed that joy was considered the most recognizable emotion (2.639, SE=0.028) and fear the least recognizable emotion (1.86, SE= 0.07) differing from all the others. The significant interaction of Emotion x Stimulus type (F 6,552 = 13.87, p˂0.00001), and relative post-hoc comparisons showed that recognizability rates were higher for emojis than faces, for all emotions except that for happiness and the neutral expressions (see Figure 2).
The significant interaction of Emotion x Sex x Stimulus type (F 6,552 = 2.3, p˂0.03), and post-hoc tests showed that women generally outperformed males in the recognition of facial expressions (especially: anger, surprise, sadness and disgust), while males outperformed women in the recognition of all emojis except for fear (see Figure 3 for means and standard deviations).
Overall, while females were better at recognizing facial expressions than men, especially surprise (p<0.04), anger (p<0.01), and sadness (p<0.01), men were better at recognizing all emojis (especially neutral, surprise and disgust, p<0.01) than women, except for fear emotion, that was better recognized by women.
Wilcoxon tests confirmed this pattern of results showing that, overall, men were significantly better than women at recognizing emojis (z= 5.99, p <0.00001) while women were better than men at recognizing human facial expressions (z= 298, p<0.00001), as shown in Figure 4. Furthermore, men were better at recognizing emojis than human facial expressions per se (z= 2.87, p <0.004).

4. Discussion

This study compared the recognizability of human facial expressions and emojis, balanced by number of emotions and number of styles/identities tested. Regardless of the sex of viewers, some emojis were recognized more clearly than facial expressions, especially anger. This result is similar to what found by Fischer and Herbert [25] contrasting facial expressions, emoji and emoticons and finding that emojis and faces were quite good in representing the associated emotions and therefore also in reducing ambiguity.
In addition, for some emotions (i.e., anger) emojis were even better than faces. It should be considered that angry and disgusted emojis, for example, have a slightly different coloring (reddish and greenish), which makes them distinctive. Quite similarly Cherbonnier and Michinov [26] comparing comprehension of real faces, emojis, and face drawings found that subjects were more accurate in detecting emotions from emojis than from other stimuli, including faces, especially for negative emotions such as disgust and fear.
Overall, in this study, happy emojis were recognized more accurately than others, while fear expressions were recognized with greater uncertainty. These data are consistent with Calvo and Lundqvist [31] reporting faster reaction times for happy and slowest reaction times for fearful expressions.
Similarly, Fischer and Herbert [25] found fastest responses for happiness and anger, followed by surprise, sadness, neutral and lastly fear. The same pattern of results was reported for real human faces [32].
These data are also in agreement with previous literature showing that children with no experience and use of social networks or smartphones recognize joyful and sad emoticons better, while they do worse in the recognition of fear and disgust [17]. The primacy of happiness expression might be related to the unique physiognomic features (e.g., the typical U-shaped mouth curvature) not present in other emotional expressions [32] or to the fact that positive expressions are more commonly displayed in social settings (e.g., the social smile), as hypothesized by Schindler and Bublatzky [33]. The same advantage for recognizing happiness was found with Emoji stimuli, also characterized by the (even more prominent) U-shaped mouth feature for smiling. The emotion of fear, on the other hand, was often confused with other emotions, particularly with surprise, as both were characterized by the arching of eyebrows and the widening of the eyes and mouth.
In this study women outperformed men in the recognition of fearful emojis. This could be partly related to the fact that fearful emojis were more obscure and difficult to be distinguished from surprise, as hits were only 57.29% in validation assessment. Similarly, in a study by Hoffmann et al. [34] it was found that women were significantly better at recognizing anger, disgust and fear (subtle emotions) than men. According to the authors, this finding was related to the women's greater ability to perceive emotions in a gestalt fashion, making quick and automatic judgements (see also [35] at this regard). Again, it was found that women were faster than men at recognizing fear, sadness and anger: they were also more accurate than men in recognizing fear [36]. In this study, both ANOVA and Wilcoxon tests showed that men were significantly better than women at recognizing emojis while women were better than men at recognizing human facial expressions Quite interestingly, men were better at recognizing emojis than human facial expressions. In general, the women’s ability to recognize human facial expressions more accurately (or faster) than men is a well corroborated notion in psychology and cognitive neuroscience [20, 37]. Further evidence has demonstrated that women, compared to men, react more strongly when viewing affective stimuli (such as IAPS) involving human beings, thus showing a higher empathic response [38-40]. Neuroimaging evidence has also been provided that face processing would engage FFA bilaterally in women, and unilaterally (i.e. only over the right hemisphere, rFFA) in men [41], thus possibly supporting a deeper/more efficient analysis of subtle facial mimicry in females. However, the presumed female superiority in the ability to comprehend emotional nonverbal cues does not seem to include emojis’ processing. Therefore, it is possible that emojis are processed differently, half-way between symbols (such as emoticons or icons) and real faces.
An interesting study by Wolf [42] observing the effect of gender in the use of emoticons in cyberspace found that to communicate with women in mixed-gender newsgroups men adopt the female standard of expressing more emotions, being therefore not less expressive as with real facial mimicry. The majority of emoticon use by females would however be lying in the meaning category of humor, and including solidarity, support, assertion of positive feelings, and thanks whereas the bulk of male emoticon use would express teasing and sarcasm. This gender difference in used emoji valence is compatible with the present pattern of results, showing how men outperformed women with negative emojis while women were better at deciphering positive emojis such as joyful or even neutral ones.

5. Study Limits and Further Directions

It should be interesting to explore a wider range of emojis to further investigate this gender effect. It would be also important to gain neuroimaging data to explore if these differences across sexes are paralleled by sex differences in the activation of visual areas devoted to face vs. emoji processing. One limit of the present study is the lack of response time data, due to the methodological paradigm, involving qualitative and quantitative assessment, but not response speed measurement (because of COVID-19 restrictions). Overall, it can be hypothesized that the unexpected male proficiency in emoji recognition might reflect a progressively greater inclination of men (namely, students) to express their emotions, particularly negative, but also amorous ones (according to other studies), through messaging, in cyberspace. In addition, it may be enlightening to compare the present pattern of results with the hyper-systemizing theory of Baron-Cohen [43] according to which males would be better at processing visuospatial information at analytic level (e.g., emojis or emoticons), and women better at empathizing with others (e.g., on fearful or sad facial expressions). These changes in social behavior should be further assessed and investigated over time.

Author Contributions

AMP conceived and planned the experiment and wrote the paper. LDN and AC prepared the stimuli and carried out the data collection. AMP performed statistical analyses and data illustration. AMP and LDN interpreted the data. All authors provided critical feedback and helped shape the research, analysis and manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of University of Milano Bicocca (CRIP, protocol number RM-2021-401) for studies involving humans.

Informed Consent Statement

Informed written consent was obtained from all subjects involved in the study.

Data Availability Statement

Anonymized data and details about preprocessing/analyses are available to colleagues upon request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANOVA Analysis of Variance
ERP Event-related potential
FFA Face Fusiform Area
fMRI functional Magnetic Resonance Imaging
IAPS International Affective Picture System
LPP Late Positive Potential
OFA Occipital face Area
SD Standard Deviation

References

  1. Bai, Q. , Dan, Q., Mu, Z.& Yang, M. A Systematic Review of Emoji: Current Research and Future Perspectives. Front Psychol. 2019, 10, 2221. [Google Scholar] [CrossRef] [PubMed]
  2. Aluja, A. , Balada, F., Blanco, E., Lucas, I. & Blanch, A. Startle reflex modulation by affective face "Emoji" pictographs. Psychol. Res. 2020, 84, 15–22. [Google Scholar] [CrossRef] [PubMed]
  3. Cherbonnier, A. , Michinov, N. The recognition of emotions beyond facial expressions: Comparing emoticons specifically designed to convey basic emotions with other modes of expression. Comput. Human Behav. 2021, 118. [Google Scholar] [CrossRef]
  4. Liao, W. , Zhang, Y., Huang, X., Xu, X. & Peng, X. “Emoji, I can feel your pain” - Neural responses to facial and emoji expressions of pain. Biol. Psychol. 2021, 163. [Google Scholar] [CrossRef]
  5. Yuasa, M., Saito, K. & Mukawa, N. Emoticons convey emotions without cognition of faces: An fMRI study. Human Factors in Computing Systems. 2006. CHI EA '06: CHI '06 Extended Abstracts on Human Factors in Computing Systems April. 1565–70. [CrossRef]
  6. Kanwisher, N. , Stanley, D., Harris, A. The fusiform face area is selective for faces not animals. Neuroreport 1999, 10, 83–87. [Google Scholar] [CrossRef] [PubMed]
  7. Chatzichristos, C. , Morante, M., Andreadis, N., Kofidis, E., Kopsinis, Y. & Theodoridis, S. Emojis influence autobiographical memory retrieval from reading words: an fMRI-based study. Plos One 2020, 15, 0234104. [Google Scholar] [CrossRef]
  8. Kim, K.W. , Lee, S.W., Choi, J., Kim, T.M., & Jeong, B. Neural correlates of text-based emoticons: a preliminary fMRI study. Brain Behav. 2016, 6, e00500. [Google Scholar] [CrossRef]
  9. Liu, j. , Harris, A. & Kanwisher, N. Perception of face parts and face configurations: an FMRI study. J. Cogn. Neurosci. 2010, 22, 203–211. [Google Scholar] [CrossRef]
  10. Huang, S.C. , Bias, R.G. & Schnyer, D. How are icons processed by the brain? Neuroimaging measures of four types of visual stimuli used in information systems. J. Assoc. Inf. Science Technol. 2015, 66, 702–720. [Google Scholar] [CrossRef]
  11. Proverbio, A.M. , Brignone V., Matarazzo S., Del Zotto M., Zani A. Gender and parental status affect the visual cortical response to infant facial expression. Neuropsychologia 2006, 44, 2987–2999. [Google Scholar] [CrossRef]
  12. Proverbio, A.M. , Zani A., Adorni R. Neural markers of a greater female responsiveness to social stimuli. BMC. Neurosci. 2008, 9, 56. [Google Scholar] [CrossRef] [PubMed]
  13. Hall, J.K. , Hutton S.B., Morgan M.J. Sex differences in scanning faces: does attention to the eyes explain female superiority in facial expression recognition? Cogn Emot 2010, 24, 629–637. [Google Scholar] [CrossRef]
  14. Thompson, A.E. , Voyer D. (2014). Sex differences in the ability to recognize non-verbal displays of emotion: a meta-analysis. Cogn. Emot. 2014, 28, 1164–1195. [Google Scholar] [CrossRef]
  15. Dimberg, U. , Lundquist, L.O. Gender differences in facial reactions to facial expressions. Biol Psychol. 1990, 30, 151–159. [Google Scholar] [CrossRef]
  16. Han S, Fan Y, Mao L. Gender difference in empathy for pain: an electrophysiological investigation. Brain Res. 2008, 1196, 85–93. [Google Scholar] [CrossRef] [PubMed]
  17. Oleszkiewicz, A. , Karwowski, M., Pisanski, K., Sorokowski, P., Sobrado, B. & Sorokowska, A. Who uses emoticons? Data from 86 702 Facebook users. Personal. Individ. Diff. 2017, 119, 289–295. [Google Scholar] [CrossRef]
  18. Jones, L. L, Wurm, L.H., Norville, G.A., Mullins, K.L. Sex differences in emoji use, familiarity, and valence. Comput. Hum. Behav. 2020, 108, 106305. [Google Scholar] [CrossRef]
  19. Kring, A. M. , & Gordon, A. H. Sex differences in emotion: Expression, experience, and physiology. J. Personal. Soc. Psychol. 1998, 74, 74–686. [Google Scholar] [CrossRef]
  20. Proverbio, A.M. Sex differences in the social brain and in social cognition. J. Neurosci. Res. 2021. [Google Scholar] [CrossRef]
  21. Chen, Z., Lu, X., Ai, W., Li, H., Mei, Q., Liu, X. Through a Gender Lens: Learning Usage Patterns of Emojis from Large-Scale Android Users. WWW 2018: The 2018 Web conference, 2018, April 23-27, Lyon. 23 April; -27.
  22. Herring, S.C. , Dainas, A.R. Receiver interpretations of emoji functions: a gender perspective. Proceedings of 1st International Workshop on Emoji Understanding and Applications in Social Media 2018. [Google Scholar]
  23. Briton, N.J. , Hall, J.A. Beliefs about female and male nonverbal communication. Sex Roles 1995, 32, 79–90. [Google Scholar] [CrossRef]
  24. Hoffmann, H. , Kessler, H., Eppel, T., Rukavina, S., Traue, H.C. Expression intensity, gender and facial emotion recognition: Women recognize only subtle facial emotions better than men. Acta Psychol. 2010, 135, 278–283. [Google Scholar] [CrossRef] [PubMed]
  25. Fischer, B. , Herbert, C. Emoji as Affective Symbols: Affective Judgments of Emoji, Emoticons, and Human Faces Varying in Emotional Content. Front. Psychol. 2021, 12, 645173. [Google Scholar] [CrossRef] [PubMed]
  26. Cherbonnier, A.; Michinov, N. The recognition of emotions beyond facial expressions: Comparing emoticons specifically designed to convey basic emotions with other modes of expression. Comput. Hum. Behav 2021, 118. [Google Scholar] [CrossRef]
  27. Ekman, P. Emotions inside out. 130 years after Darwin’s “The Expression of the Emotions in Man and Animal”. Ann. N. Y. Acad. Sci. 2003, 1000, 1–6. [Google Scholar] [CrossRef]
  28. Carbon, C.C. Wearing Face Masks Strongly Confuses Counterparts in Reading Emotions. Front. Psychol. 2020, 11, 566886. [Google Scholar] [CrossRef] [PubMed]
  29. Aviezer, H., Hassin. Angry, disgusted, or afraid? Studies on the malleability of emotion perception. Psychol Sci. 2008, 19, 724–732. [Google Scholar] [CrossRef] [PubMed]
  30. Derntl, B. , Seidel, E.M., Kainz, E., Carbon, C.C. Recognition of emotional expressions is affected by inversion and presentation time. Perception. 2009, 38, 1849–1862. [Google Scholar] [CrossRef]
  31. Calvo, M. G. , Lundqvist, D. Facial expressions of emotion (KDEF): Identification under different display-duration conditions. Behav. Res. Methods. 2008, 40, 109–115. [Google Scholar] [CrossRef]
  32. Sun, A. , Li, Y., Huang, Y.-M., Li, Q. & Lu, G. (2018). Facial expression recognition using optimized active regions. Hum.-Centric Comput. Inform. Scienc. 2018, 8, 1–24. [Google Scholar] [CrossRef]
  33. Schindler, S. & Bublatzky, F. Attention and emotion: An integrative review of emotional face processing as a function of attention. Cortex. 2020, 30, 362–386. [Google Scholar] [CrossRef]
  34. Hoffmann, H. , Kessler, H., Eppel, T., Rukavina, S., Traue, H.C. Expression intensity, gender and facial emotion recognition: Women recognize only subtle facial emotions better than men. Acta Psychol. 2010, 135, 278–283. [Google Scholar] [CrossRef]
  35. Hall, J.A. , Matsumoto, D. Gender differences in judgments of multiple emotions from facial expressions. Emotion. 2004, 4, 201–206. [Google Scholar] [CrossRef] [PubMed]
  36. Kapitanović, A. , Tokić, A., Šimić, N. Differences in the recognition of sadness, anger, and fear in facial expressions: the role of the observer and model gender. Arh Hig Rada Toksikol. 2023, 73, 308–313. [Google Scholar] [CrossRef] [PubMed]
  37. Wingenbach, T.S.H. , Ashwin, C., Brosnan, M. Sex differences in facial emotion recognition across varying expression intensity levels from videos. PLoS One. 2018, 13, e0190634. [Google Scholar] [CrossRef] [PubMed]
  38. Klein, S. , Smolka, M.N., Wrase, J., et al. The influence of gender and emotional valence of visual cues on fMRI activation in humans. Pharmacopsychiat. 2003, 36, 5191–5194. [Google Scholar] [CrossRef]
  39. Gard, M.G. , Kring, A.M. Sex differences in the time course of emotion. Emotion, 2007, 7, 429–437. [Google Scholar] [CrossRef]
  40. Proverbio, A.M. , Adorni, R., Zani, A., Trestianu, L. Sex differences in the brain response to affective scenes with or without humans. Neuropsychologia. 2009, 47, 2374–2388. [Google Scholar] [CrossRef]
  41. Proverbio, A.M. Sexual dimorphism in hemispheric processing of faces in humans: A meta-analysis of 817 cases. Soc. Cogn. Affect. Neurosci. 2021, 16, 1023–1035. [Google Scholar] [CrossRef]
  42. Wolf, A. Emotional Expression Online: Gender Differences in Emoticon Use. Cyberpsychol. Behav. 2000, 3, 827–833. [Google Scholar] [CrossRef]
  43. Baron-Cohen, S. The extreme male brain theory of autism. Trends Cogn. Sci. 2002, 6, 248–254. [Google Scholar] [CrossRef]
Figure 1. Examples of facial expressions and emojis used to illustrate the 7 affective states, in 4 different “identities”: 1 female, 1 male, and 2 different emoji styles.
Figure 1. Examples of facial expressions and emojis used to illustrate the 7 affective states, in 4 different “identities”: 1 female, 1 male, and 2 different emoji styles.
Preprints 67732 g001
Figure 2. Mean recognizability scores (with standard deviations) relative to the 7 expressions illustrated by human faces and emoji.
Figure 2. Mean recognizability scores (with standard deviations) relative to the 7 expressions illustrated by human faces and emoji.
Preprints 67732 g002
Figure 3. Mean recognizability scores (with standard deviations) measured in male and female participants in response to the seven emotions expressed by human faces and emojis.
Figure 3. Mean recognizability scores (with standard deviations) measured in male and female participants in response to the seven emotions expressed by human faces and emojis.
Preprints 67732 g003
Figure 4. Distribution of individual performance as a function of stimulus type and sex of viewers. On the y-axis are visible recognizability scores, while on the x-axis are subjects’ responses. On the upper part of the figure, the scores are grouped as a function of sex of viewers. M= male, F= Female participants.
Figure 4. Distribution of individual performance as a function of stimulus type and sex of viewers. On the y-axis are visible recognizability scores, while on the x-axis are subjects’ responses. On the upper part of the figure, the scores are grouped as a function of sex of viewers. M= male, F= Female participants.
Preprints 67732 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Alerts
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2025 MDPI (Basel, Switzerland) unless otherwise stated