The business of a sentence is not only to describe some state of affairs but also to perform other kinds of speech acts like ordering, suggesting, asking, etc. Understanding the kind of action performed by a speaker who utters a sentence is a multimodal process which involves the computing of verbal and non-verbal information. This work aims at investigating if the understanding of a speech act is affected by the gender of the actor that produces the utterance in combination with a certain facial expression. Experimental data collected show that, as compared to men, women are less likely to be perceived as performers of orders and are more likely to be perceived as performers of questions. This result reveals a gender bias which reflects a process of women’s subordination according to which women are hardly considered as holding the hierarchical social position required for the correct execution of an order
AllwoodJ.CerratoL.JokinenK.NavarrettaC. & PaggioP. (2005). The MUMIN annotation scheme for feedback, turn management and sequencing. In Gothenburg papers in Theoretical Linguistics 92: Proceedings from The Second Nordic Conference on Multimodal Communication pp. 91–109. Göteborg, Sweden: Göteborg University.
DomaneschiF.PassarelliM. & ChiorriC. (2017). Facial expressions and speech acts: Experimental evidences on the role of the upper face as an illocutionary force indicating device in language comprehension. Cognitive Processing. Advance online publication.
EaglyA.H.WoodW. & DiekmanA. (2000). Social role theory of sex differences and similarities: A current appraisal. In T.Eckes & H.M.Trautner (Eds.) The developmental social psychology of gender (pp. 123–174). Washington, DC: Taylor & Francis.
EspositoA.EspositoA.M.MartoneR.MüllerV.ScarpettaG. (Eds.) (2010). Towards Autonomous Adaptive and Context-Aware Multimodal Interfaces: Theoretical and Practical Issues. New York, NY & London, UK: Springer.
PelachaudC.CarofiglioV.De CarolisB.de RosisF. & PoggiI. (2002, July). Embodied contextual agent in information delivering application. In Proceedings of the first international joint conference on Autonomous agents and multiagent systems: Part 2 (pp. 758–765). New York, NY: ACM.
PelachaudC.PrevostS. (1994). Sight and sound: Generating facial expressions and spoken intonation from context. In Proceedings of the 2nd Workshop on Speech Synthesis (216–219). New York, NY: New Paltz.