• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 1
  • 1
  • Tagged with
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

以參數化程序產生具情緒表達能力之3D肢體動畫 / Designing Parameterized Procedures for Real-Time 3D Figure Animation with Affective Expression

林岳黌, Lin, Yueh Hung Unknown Date (has links)
人或擬人生物佔眾多動畫主題的大部份,而要使人物動畫看起來栩栩如生,除了適當的臉部表情外,能夠傳達人物情感的肢體動作更是不可或缺。本研究的目標在於以電腦程序產生帶有情緒成份之人物肢體動畫。此目標包含兩個子目標:第一、設計以人類肢體動作為目標的參數化程序,以提高程序之重用性及泛用性,降低製作程序式動畫的成本,並製作出以關鍵格為基礎之人物動畫;第二、將我們製作的動畫加入風格,並以心理學實驗驗證人物的動作和情緒的關聯。我們的實驗先將風格套用於走路動作上,來證明我們操弄風格的方法符合大眾的認知。再者,我們針對走路動作嘗試以實驗找出風格和情緒的對映,再將此對映關係套用到具特定情緒意涵的動作上,以證明這些對映有助於情緒的表達。 / Human or human-like creatures are the main subjects of computer animations. In addition to facial expression, body gestures and motions are also indispensi-ble components for realistic character animation. The goal of this research is to create emotional character animation with computer procedures. This goal may contain two subgoals: first, we aim to design parameterizable animation proce-dures for human body motions in order to reduce the cost of producing key-frame based character animations with improved generality and reusability; second, we incorporate style into procedural animation and validate the relation model between emotion and motion with psychology experiments. We first ap-plied different styles into the walking motion and conducted experiments to see if the participants can agree with the way that we manipulate the style parame-ters. Furthermore, for the walking motion, we conduct experiments to find the mapping from the emotion parameters to the style parameters. Then we applied this mapping to emotion-specific motions to see if the animations perceived by the users can be further enhanced.

Page generated in 0.1198 seconds