• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 71
  • 35
  • 16
  • 8
  • 5
  • 4
  • 2
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 176
  • 176
  • 59
  • 36
  • 32
  • 31
  • 30
  • 25
  • 22
  • 22
  • 21
  • 18
  • 18
  • 17
  • 16
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
11

Effects of Emotional Expressions on Eye Gaze Discrimination and Attentional Cuing

Lee, Daniel Hyuk-Joon 15 February 2010 (has links)
Recent evidence has shown that our emotional facial expressions evolved to functionally benefit the expression’s sender, in particular fear increasing and disgust decreasing sensory acquisition. Using schematic eyes only that lack emotional content, but taken from actual participant fear and disgust expressions, we examined the functional action resonance hypothesis that adaptive benefits are also conferred to the expression’s receiver. Participants’ eye gaze discrimination was enhanced when viewing wider, “fear” eyes versus narrower, “disgust” eyes (Experiment 1). Using a gaze cuing paradigm, task facilitation by way of faster responses to target was found when viewing wider versus narrower eyes (Experiment 2). Contrary to our hypothesis, a null attention modulation for wider versus narrower eyes was found (Experiments 2 and 3). Nonetheless, the evidence is argued for the functional action resonance hypothesis.
12

The role of affective information in context on the judgment of facial expression: in what situations are North Americans influenced by contextual information?

Ito, Kenichi 11 1900 (has links)
Research in cultural psychology suggests that East Asians are more likely than North Americans to be sensitive to contextual information. By contrast, much evidence suggests that even North Americans judgments are influenced by affective priming information, the effect of which can be seen as another type of contextual cue. However, the magnitude of such priming effect has not been tested in a cross-cultural context. Using the methodology of the affective priming paradigm, we conducted two studies, in which we manipulated (a) the timing of priming information (simultaneous vs. sequential) and (b) the type of affective information (background landscape vs. background human figures), in which European Canadians and Japanese judged either happy or sad facial expressions in the focal area of the scene. The results indicate that the two cultural groups are similar when contextual information is salient, but only Japanese remain sensitive to context with subtle cues.
13

The empathy fillip : can training in microexpressions of emotion enhance empathic accuracy?

Eyles, Kieren January 2016 (has links)
Empathy is a central concern in the counselling process. Though much researched, and broadly commented upon, empathy is still largely understood through the words within a client-counsellor interaction. This semantic focus continues despite converging lines of evidence that suggest other elements of an interaction – for example body language – may be involved in the communication of empathy. In this thesis, the foundations of empathy are examined, focusing on empathy’s professional instantiation. These foundations are then related to the idea that the face, and its ability to express emotion, are an important part of the empathic process. What follows is an experiment testing 60 participants. This was a between groups design, with participants assigned to two even groups; one group receiving training in how emotion appears on the face: using the training program eMETT; the other reading a passage on empathy. Following the intervention, hypothesised group differences were assessed using the following analyses. Firstly, an Independent sample T-test, compared group means on the Ickes Empathic Accuracy paradigm, the measure of empathy used. Secondly, a further Independent sample T-test assessed the effect of eMETT training. Thirdly, an ANCOVA, evaluated whether the obtained results may have been confounded by age difference between the experimental groups. Finally a correlational analyse tested for any relationship between baseline and outcome measures. The hypothesis tested stated: training in facial expressions of emotion will enhance counsellors’ empathic accuracy; a hypothesis for which positive evidence was shown. The implications of this evidence suggest efficacy of the eMETT training to enhance empathic accuracy, though this is qualified through critical examination of the experimental method. Suggestions for refinement of this method are discussed.
14

Automatic facial expression analysis

Baltrušaitis, Tadas January 2014 (has links)
Humans spend a large amount of their time interacting with computers of one type or another. However, computers are emotionally blind and indifferent to the affective states of their users. Human-computer interaction which does not consider emotions, ignores a whole channel of available information. Faces contain a large portion of our emotionally expressive behaviour. We use facial expressions to display our emotional states and to manage our interactions. Furthermore, we express and read emotions in faces effortlessly. However, automatic understanding of facial expressions is a very difficult task computationally, especially in the presence of highly variable pose, expression and illumination. My work furthers the field of automatic facial expression tracking by tackling these issues, bringing emotionally aware computing closer to reality. Firstly, I present an in-depth analysis of the Constrained Local Model (CLM) for facial expression and head pose tracking. I propose a number of extensions that make location of facial features more accurate. Secondly, I introduce a 3D Constrained Local Model (CLM-Z) which takes full advantage of depth information available from various range scanners. CLM-Z is robust to changes in illumination and shows better facial tracking performance. Thirdly, I present the Constrained Local Neural Field (CLNF), a novel instance of CLM that deals with the issues of facial tracking in complex scenes. It achieves this through the use of a novel landmark detector and a novel CLM fitting algorithm. CLNF outperforms state-of-the-art models for facial tracking in presence of difficult illumination and varying pose. Lastly, I demonstrate how tracked facial expressions can be used for emotion inference from videos. I also show how the tools developed for facial tracking can be applied to emotion inference in music.
15

Seven- to 11-Year-Olds' Developing Ability to Recognize Natural Facial Expressions of Basic Emotions

Kang, K., Anthoney, L., Mitchell, Peter 04 June 2020 (has links)
Yes / Being able to recognize facial expressions of basic emotions is of great importance to social development. However, we still know surprisingly little about children’s developing ability to interpret emotions that are expressed dynamically, naturally, and subtly, despite real-life expressions having such appearance in the vast majority of cases. The current research employs a new technique of capturing dynamic, subtly expressed natural emotional displays (happy, sad, angry, shocked, and disgusted). Children aged 7, 9, and 11 years (and adults) were systematically able to discriminate each emotional display from alternatives in a five-way choice. Children were most accurate in identifying the expression of happiness and were also relatively accurate in identifying the expression of sadness; they were far less accurate than adults in identifying shocked and disgusted. Children who performed well academically also tended to be the most accurate in recognizing expressions, and this relationship maintained independently of chronological age. Generally, the findings testify to a well-developed ability to recognize very subtle naturally occurring expressions of emotions.
16

Animals in animation : Anthropomorphised facial expressions and the uncanny valley: does a stylized or realistic 3D animal become uncanny when applying anthropomorphised or realistic facial expressions to it?: Does it change depending on the expression?

Frick, Gustav, Malinen, Lovisa, Ryan, Victoria January 2022 (has links)
People have a tendency to apply human characteristics to animals, i.e anthropomorphisation. With this in mind, along with the ever increasing number of CGI animals in animated media today, this paper examines whether or not perceived eeriness of a 3D cat model increases when there is a mismatch between realism/stylisation in the style of the model and the animation presented. The familiarity a person has towards something increases as that something becomes more human-like, but at a certain point, familiarity dips into what is known as the Uncanny Valley. In this study we research whether or not this phenomenon is exacerbated when a mismatch between a realistic model and stylised animation is applied, or vice versa, using both quantitative and qualitative data. Our results indicate that there is in fact an increase in uncanniness when a mismatch is present, especially on the realistic model, though future work is required tomake a definitive link. / <p>Det finns övrigt digitalt material (t.ex. film-, bild- eller ljudfiler) eller modeller/artefakter tillhörande examensarbetet som ska skickas till arkivet.</p>
17

Protecting your interviewer's face : how job seekers perceive face threat in a job interview

Howell, Catherine Ray, 1985- 28 October 2010 (has links)
The interview is an important component of the selection process for employment and is one of the initial presentations of self by the applicant to the interviewer. As an extension of a study by Wilson, Aleman, and Leatham (1998), this study used politeness theory to investigate perception of face threat in the context of a job interview, specifically, when making requests and giving advice. This study predicted that jobseekers perceive an act as a greater threat to an interviewer’s negative face (appealing to interviewer’s autonomy) when making a request than when giving advice. Secondly, the study predicted that job seekers would perceive an act as greater threat to the interviewer’s positive face (appealing to the interviewer’s desire for approval) when giving advice or recommendations than when making a request. Both hypotheses were supported and other related interests such as acceptability of the act and likelihood of getting the job were also investigated. / text
18

Sensitivity to Emotion Specified in Facial Expressions and the Impact of Aging and Alzheimer's Disease

McLellan, Tracey Lee January 2008 (has links)
This thesis describes a program of research that investigated the sensitivity of healthy young adults, healthy older adults and individuals with Alzheimer’s disease (AD) to happiness, sadness and fear emotion specified in facial expressions. In particular, the research investigated the sensitivity of these individuals to the distinctions between spontaneous expressions of emotional experience (genuine expressions) and deliberate, simulated expressions of emotional experience (posed expressions). The specific focus was to examine whether aging and/or AD effects sensitivity to the target emotions. Emotion-categorization and priming tasks were completed by all participants. The tasks employed an original set of cologically valid facial displays generated specifically for the present research. The categorization task (Experiments 1a, 2a, 3a, 4a) required participants to judge whether targets were, or were not showing and feeling each target emotion. The results showed that all 3 groups identified a genuine expression as both showing and feeling the target emotion whilst a posed expression was identified more frequently as showing than feeling the emotion. Signal detection analysis demonstrated that all 3 groups were sensitive to the expression of emotion, reliably differentiating expressions of experienced emotion (genuine expression) from expressions unrelated to emotional experience (posed and neutral expressions). In addition, both healthy young and older adults could reliably differentiate between posed and genuine expressions of happiness and sadness, whereas, individuals with AD could not. Sensitivity to emotion specified in facial expressions was found to be emotion specific and to be independent of both the level of general cognitive functioning and of specific cognitive functions. The priming task (Experiments 1b, 2b, 3b,4b) employed the facial expressions as primes in a word valence task in order to investigate spontaneous attention to facial expression. Healthy young adults only showed an emotion-congruency priming effect for genuine expressions. Healthy older adults and individuals with AD showed no priming effects. Results are discussed in terms of the understanding of the recognition of emotional states in others and the impact of aging and AD on the recognition of emotional states. Consideration is given to how these findings might influence the care and management of individuals with AD.
19

Emotional Sophistication: Studies of Facial Expressions in Games

Rossi, Filippo January 2012 (has links)
Decision-making is a complex process. Monetary incentives constitute one of the forces driving it, however the motivational space of decision-makers is much broader. We care about other people, we experience emotional reactions, and sometimes we make mistakes. Such social motivations (Sanfey, 2007) drive our own decisions, as well as affect our beliefs about what motivates others' decisions. Behavioral and brain sciences have started addressing the role of social motivations in economic games (Camerer, 2004; Glimcher et al., 2009), however several aspects of social decisions, such as the process of thinking about others' emotional states - emotional sophistication - have been rarely investigated. The goal of this project is to use automatic measurements of dynamic facial expressions to investigate non-monetary motivations and emotional sophistication. The core of our approach is to use state-of-the-art computer vision techniques to extract facial actions from videos in real-time (based on the Facial Action Coding System of Ekman and Friesen (1978)), while participants are playing economic games. We will use powerful statistical machine learning techniques to make inferences about participants internal emotional states during these interactions. These inferences will be used (a) to predict behavior; (b) to explain why a decision is made in terms of the hidden forces driving it; and (c) to investigate the ways in which people construct their beliefs about other people's future actions. The contributions of this targeted interdisciplinary project are threefold. First, it develops new methodologies to study decision processes. Second, it uses these methods to test hypotheses about the role of first order beliefs about social motivations. Finally, our statistical approach sets the ground for "affectively aware" systems, that can use facial expressions to assess the internal states of their users, thus improving human-machine interactions.
20

Effects of Gender and Self-Monitoring on Observer Accuracy in Decoding Affect Displays

Spencer, R. Keith (Raymond Keith) 12 1900 (has links)
This study examined gender and self-monitoring as separate and interacting variables predicting judgmental accuracy on the part of observers of facial expressions of emotional categories. The main and interaction effects failed to reach significant levels during the preliminary analysis. However, post hoc analyses demonstrated a significant encoder sex variable. Female encoders of emotion were judged more accurately by both sexes. Additionally, when the stimulus was limited to female enactments of emotional categories, the hypothesized main and interaction effects reached significant F levels. This study utilized 100 observers and 10 encoders of seven emotional categories. Methodological considerations and alternatives are examined at length.

Page generated in 0.073 seconds