A New Generation of Synthetic Actors: the Real-time and Interactive Perceptive Actors*

Size: px
Start display at page:

Download "A New Generation of Synthetic Actors: the Real-time and Interactive Perceptive Actors*"

Transcription

1 A New Generation of Synthetic Actors: the Real-time and Interactive Perceptive Actors* Daniel Thalmann Computer Graphics Lab (LIG), Swiss Federal Institute of Technology (EPFL) CH-1015 Lausanne, Switzerland In this paper, we show the evolution from the earliest frame-by-frame synthetic actors to real-time perceptive actors living in distributed virtual environments or even in the real world. We first compare requirements and methods for frame-by-frame and real-time actors. Then, we introduce a new classification of synthetic actors with four categories: participants, guided actors, autonomous actors, and interactive perceptive actors. Actors of this last category are not only aware of other actors and real people, but they are also able to communicate with them. keywords: participant, guided actor, autonomous actor, interactive perceptive actor, interactive tennis 1 Synthetic Actors in Movies Since a few years, the terms of virtual, synthetic or digital actors have been extensively used in order to describe very different kinds of entities. Because of the lack of power of computers and workstations, synthetic actors were typically generated frame-by-frame for movie production. For example, in 1982, we produced the first short feature film Dream Flight [1] involving Hippie a wireframe synthetic actor. Hippie was a complete procedural actor. His motion was decided by the animators and programmed using the MIRA Pascal-based object-oriented language [2]. But, the term actor is well-known in the world of cinema and everybody knows it. A Hollywood-like actor is first a human personality. That is the reason, we only introduced the term synthetic actor when recreating Marilyn Monroe and Humphrey Bogart for the film Rendez-vous à Montréal [3]. The motion of Marilyn and Bogey were created using the Human Factory software, with mainly keyframe animation for the body motion. A specific facial animation software was developed and body deformations were created using JLD operators. Today, most movies involving synthetic actors and 3D characters are still now produced with a similar approach using commercial software. Although the current software offer some facilities like inverse kinematics and dynamics, keyframe is still very popular. Actors are moved by animators, who set body positions in keyframes, let the computer interpolate between the frames, render the animation frame by frame, store the resulting data, and, finally, record the frames on tape or film. A few companies also use motion capture for the production of computer-generated sequences with 3D characters. Performance animation or motion capture consists of measurement and recording of direct actions of a real person or animal for immediate or delayed analysis and playback. The technique involves mapping of measurements onto the motion of the digital character. Maiocchi [4] gives more details about performance animation. We may distinguish three kinds of systems: mechanical, magnetic, and optical. Motion capture methods offer advantages and disadvantages. Let us consider the case of human walking. A walking motion may be recorded and then applied to a computer-generated 3D character. It will provide a very good motion, because it comes directly from reality. However, motion capture does not bring any really new concept to animation methodology. For any new motion, it is necessary to record the reality again. 2 The Role of Real-time Actors in Multimedia and VR With the new developments of digital and interactive television [5] and multimedia products, the viewer will be more and more able to interact with programs and this will lead to individual programs for each viewer. There is a need for systems that provide designers with the capability for embedding real-time simulated humans in games, multimedia titles and film animations. The ability to embed the viewer in a dramatic situation created by the behaviour of other, simulated, digital actors will add a new dimension to existing simulation-based products for the education and entertainment on interactive TV. In the games market, convincing simulated humans have been identified by the industry as a way of giving a fresh appearance to existing games and enabling new kinds of game to be produced. Finally, in Virtual Reality, representing participants by a virtual actor is an important factor for presence. This becomes even *Published in Proc. Paciphic Graphics '96, Taipeh, Taiwan, 1996, pp

2 more important in multi-user environments, where effective interaction among participants is a contributing factor to presence. This is the Self-representation in the Virtual World. Even with limited sensor information, a virtual human frame can be constructed in the Virtual World, that reflects the activities of the real body. Slater and Usoh [6] indicates that such a body, even if crude, heightens the sense of presence. A few companies have started to move 3D characters in real time, much as a puppeteer controls a puppet. However, motion capture techniques are not appropriate especially in real-time simulation activities, where the situation and actions of people cannot be predicted ahead of time, and in dangerous situations, where one cannot involve a human actor. Now, the advent of the hardware new technologies and the integration of techniques of Computer Animation, Artificial Intelligence, and Artificial Life allow to create truly synthetic actors. In this paper, synthetic (virtual or digital) actors will be considered as human-like actors completely generated by computer. 3 A Comparison between Frame-by-Frame and Real-Time Actors In this Section, we try to identify the various techniques appropriate for the development of frame-byframe and real-time actors. We arbitrary classify these methods in: modeling and deformations motion control: skeleton animation and facial animation realistic aspects: clothes, hair, textures, and shadows actors in Virtual Environments and Real World Table 1 compares the methods for both types of actors: frame by frame and real-time. frame-by-frame real-time surface modeling no limitations on the limitations on the number of polygons complexity deformations may be calculated using requires fast transformations, e.g. based on metaballs, FFD, splines cross sections skeleton animation any method may be used requires real-time processing which may prevent to use expensive methods based on inverse dynamics or control theory locomotion any model/method may be dynamic models may be too CPU intensive used: motion capture, kinematics, dynamics, biomechanics grasping complex models may be used kinematics and heuristic methods should be facial animation including inverse dynamics complex models may be used, including muscles with finite elements used simplified models should be used; limitations of the facial deformations in virtual not applicable no specific problems except limitations for environments immersive applications in the real world may be composed off line using a videocomposer requires real-time blending using a powerful video card like Sirius shadow true shadows may be computed only shadows of approximate primitives may be calculated clothes autonomous clothes calculated texture mapping using mechanical models skin model with wrinkles texture mapping hair individual hair possible only a polygonal shape with possible texture may be applied Table 1. Comparison between frame-by-frame and real-time actors Autonomous Virtual Humans

3 3.1 Actor Modeling and Deformations In order to create a well-known personality like Marilyn, this is essential to create an accurate threedimensional face. There are two main ways to do it: 3D digitizing and sculpting [7]. For constructing and animating realistic human bodies, the most effective approach is the multi-layered model [8, 9]. It contains a skeleton layer, intermediate layers which simulate the physical behavior of muscle, bone, fat tissue, etc., and a skin layer. Jianhua and Thalmann [10] describe a multi-layered approach. Metaballs are employed to simulate the gross behavior of bone, muscle, and fat tissue. They are attached to the proximal joints of the skeleton, arranged in an anatomically-based approximation. The skin surfaces are automatically constructed using cross-sectional sampling. The method combines the advantages of implicit, parametric and polygonal surface representation, producing realistic body deformations. Figure 1 shows a body built using this method. Fig. 1. Body built using metaballs and B-splines Although this method is fast (2-3 seconds per frame), it is not real-time. Recently, Thalmann et al. [11] described a derived approach based on contours deformation. First, they compute these contours points in a given position of the body. At this point, they don t compute anymore B-spline patches and triangulate these surfaces: contours points are used directly for body triangles mesh. Each contour belongs to one body part, and the body parts are used to construct the final 3D surface. Then, during real-time animation, contours are directly deformed still using joint angle driven configuration and boundary driven configuration methods. By this way, contours computation from implicit surfaces are avoided, which is, with B-spline patches triangulation, the most time-consuming operation. The implementation is based on the Performer library, the real-time graphics toolkit from Silicon Graphics. In real-time applications, an efficient way to reduce computing time is to manage different levels-of-detail (LOD) for one object: considering the distance between the camera and our object, we display this object in the best resolution for this distance. The body envelope is one unique surface composed of eleven triangles meshes. Dimensions of each triangle mesh is given by the number of contours and the number of points per contour. By modify these numbers, it is possible to get different LODs for each body part and thus for all the body. 3.2 Motion Control Body motion control is normally performed by animating a skeleton. Using geometric techniques, the skeleton is locally controlled and defined in terms of coordinates, angles, velocities, or accelerations. The simplest approach is motion capture, but key frame animation is still another popular technique in which the animator explicitly specifies the kinematics by supplying keyframes values whose "in-between" frames are interpolated by the computer. Inverse kinematics [12] is a technique coming from robotics, where the motion of links of a chain is computed from the end link trajectory. Geometric methods are efficient and Autonomous Virtual Humans - 3 -

4 may be easily performed in real-time. However, even if keyframe animation is in real-time, the design of the keys should be prepared in advance. In physics-based methods, the animator provides physical data and the motion is obtained by solving the dynamic equations. Motion is globally controlled. We may distinguish methods based on parameter adjustment [13] and constraint-based methods, where the animator states in terms of constraints the properties the model is supposed to have, without needing to adjust parameters. For example, Witkin and Kass [14] propose the concept of Spacetime Constraints, for creating character animation by solving constrained optimization. Cohen [15] takes this concept further and uses spacetime window to control the animation interactively. Some physics-based methods may be very time-consuming. Concerning facial animation, for frame by frame actors, complex models may be used including muscles with finite elements. For real-time actors, simplified geometric models [16] should be used with limitations of the facial deformations. 3.3 Realistic Aspects In order to improve the appearance of synthetic actors, hairstyles, skin, and clothes should be designed for them. Recently, several authors have proposed algorithms for representing realistic hair, but these algorithms are far from real-time because of the great number of geometrical primitives involved, the potential diversity of the curvature of each strand of hair, and the light and self-shadowing effects. For realtime processing, hair should be modeled as a simple polygonal mesh. For skin rendering in real-time, texture mapping may be also used although this is more difficult to produce good results. For a frame-byframe production, it is possible to produce wrinkles using a skin model [17]. Dressing synthetic actors with complex clothes is also a challenge. It is somewhat difficult to realistically animate complex objects consisting of many surface panels like trousers or jackets without proper dynamic constraints. Problems include seaming the surface panels together, attaching them to other rigid objects, and calculating collision responses when they self-collide or collide with rigid objects. However, the problem was solved recently [18, 19] but not for real-time actors. For real-time processing, we can apply different textures on each body part. Using hardware texture mapping, performance is the same than without texture, but enable texturing is a good way to increase surface realism. Figure 2 shows an example. Fig. 2. Body surface with texturing Autonomous Virtual Humans

5 Another important aspect to make the scene more realistic is the production of shadows. True and accurate shadows are generally produced by ray tracing. Simple shadows of approximate shapes may be easily projected on the floor for real-time processing. 3.4 Actors in Virtual Environments and the Real World Virtual Environments are real-time by definition and there is no sense to consider the problem of frameby-frame actors in this case. However, we should mention the limitation in the complexity of actor models for immersive environments. Although this is possible to consider up to 8000 polygons for a model rendered on a Onyx, the use of most HMDs should drastically limit the number of polygons. For creating scenes involving virtual actors in the real world, we should really take into account the real world during the generation of the images by the computer. For example, consider a virtual actor passing behind a real tree: for some images of the actor, part of the body should be hidden. For more realism, the shadow of the actor should be cast on the real floor. This means that the computer-generated images are dependent on the real world. One way of solving these problems is to create virtual objects similar to the real ones and a virtual camera corresponding to the real camera. However, this correspondence is generally hard to establish. The process is tedious and requires composition stages [20]. In order to produce real-time virtual actors in a real world, it requires special equipment and the real world should be a specific world well-known in advance. For example, we could mention interactive video environments like the ALIVE system [21]. In this environment virtual actors and the user can see each other --- users can see the actors on the video screen, and the actors can see users through a computer vision system. An image of the user appears on the video screen, effecting a type of magic mirror, in which users see themselves in a different world through the use of a simulated mirror. Before other processing can occur, the vision system must isolate the figure of the user from the background using low-level image processing techniques to detect differences in the scene, and use connectedcomponents analysis routines to extract objects. We also developed a system for real-time insertion of synthetic actors in our lab. The system is working on an Onyx with the Sirius video card. With this system, we are able to display a synthetic real-time Marilyn in one part of our lab. This Marilyn could be together with people of the lab. She has a simplified shadow on the real floor and may be hidden by specific objects in the lab. Figure 3 shows an example. Fig. 3. Marilyn visits our lab in real-time 4 A New Classification of Synthetic Actors 4.1 Previous Classifications There are a lot of methods for controlling motion of synthetic actors. For example, Zeltzer [22] classifies animation systems as being either guiding, animator-level or task-level systems. Magnenat Thalmann and Thalmann [23] propose a classification of computer animation scenes involving synthetic actors both according to the method of controlling motion and according to the kinds of interactions the actors have. Autonomous Virtual Humans - 5 -

6 The nature of privileged information for the motion control of actors falls into three categories: geometric, physical and behavioral, giving rise to three corresponding categories of motion control method. The first approach corresponds to methods heavily relied upon by the animator: performance animation, shape transformation, parametric keyframe animation. Animated objects are locally controlled. Methods are normally driven by geometric data. The second way guarantees a realistic motion by using physical laws, especially dynamic simulation. The animator should provide physical data corresponding to the complete definition of a motion. The motion is obtained by the dynamic equations of motion relating the forces, torques, constraints and the mass distribution of objects. The third type of animation is called behavioral animation and takes into account the relationship between each object and the other objects. Moreover the control of animation may be performed at a task level, but we may also consider the animated objects as autonomous creatures. 4.2 The Proposed Classification We propose a new classification of synthetic actors with four types: participants guided actors autonomous actors interactive perceptive actors Table 2 compares the various types of actors for several types of actions. participants guided actors autonomous actors walking on flat only possible requires a walking requires a terrain with full sensors motor and walking motor procedure to input velocity and step length walking on non only possible no current model difficult to have a flat terrain with full sensors general walking motor; genetic algorithms are promising grasping possible with requires a possible with gloves, but hard procedure of automatic to implement inverse kinematics procedure feedback; use of fine tuning can help obstacle avoidance not possible could be vision-based implemented using navigation algorithms based on graph theory gesture recognition not provided not provided generally not provided communication communication communication is limited with actors is not computergenerategenerated not computer- communication communication corresponds to a corresponds to a generally not with user communication communication provided between users between users interactive perceptive actors requires a walking motor difficult to have a general walking motor; genetic algorithms are promising possible with automatic procedure vision-based navigation should be provided perception-based communication verbal and non-verbal easy through the participant corresponding to the user Table 2. Types of actors for several types of actions 4.3 Participants or Virtual Actors as Self Representations of the User The virtual actor is required to have a natural-looking body and be animated correlated to the actual body. The technique, already mentioned in Section 2, may be called real-time rotoscopy method [24] and consists of recording input data from a VR device in real-time allowing to apply at the same time the same data to the virtual actor on the screen. For example, when the user opens the fingers 3 centimeters Autonomous Virtual Humans

7 using a CyberGlove, the hand on the screen does exactly the same. As shown in Figure 4, a popular way of animating such an actor is the use of sensors like the Flock of Birds. 4.4 Guided Actors Guided actors are actors which are driven by the user but which do not correspond directly to the user motion. They are based on the concept of real-time direct metaphor [24]. To understand the concept, let us have an example of traditional metaphor: the puppet control. A puppet may be defined as a doll with jointed limbs moved by wires or strings. Human fingers are used to drive the motion of the puppet. The best example of actor guidance is guided navigation like implemented in VLNET [25] (see Section 5.3 and Figure 5). The animator participant uses the input devices to update the position of the virtual actor. Fig. 4. Animating a Virtual Actor using a Flock of Birds Fig. 5. A guided actor in the VLNET system This local control is used by computing the incremental change in the actor position, and estimating the rotation and velocity of the center of body. The walking motor [26] uses the instantaneous velocity of motion, to compute the walking cycle length and time, by which it computes the necessary joint angles. The sensor information for walking can be obtained from various types of input devices like DataGlove, SpaceBall, or virtual treadmill as proposed by Slater et al. [27]. Autonomous Virtual Humans - 7 -

8 4.5 Autonomous Actors Autonomy through Virtual Sensors Autonomous actors are able to have a behavior, which means they must have a manner of conducting themselves. Behavior is not only reacting to the environment but should also include the flow of information by which the environment acts on the living creature as well as the way the creature codes and uses this information. If we consider the synthetic environment as made of 3D geometric shapes, one solution to this problem is to give the actor access to the exact position of each object in the complete environment database corresponding to the synthetic world. This solution could work for a very "small world", but it becomes impracticable when the number of objects increases. Moreover, this approach does not correspond to reality where people do not have knowledge about the complete environment. In a typical behavioral animation [28] scene, the actor should perceive the objects and the other actors in the environment through visual, tactile (including haptic) and auditory sensors. These virtual sensors [29] should be used as a basis for implementing everyday human behaviour such as visually directed locomotion, handling objects, and responding to sounds and utterances. Based on the perceived information, the actor s behavioral mechanism will determine the actions he will perform. Actions may be at several degrees of complexity. An actor may simply evolve in his environment or he may interact with this environment or even communicate with other actors. In this latter case, we will consider the actor as a interactive perceptive actor Virtual Vision Although the use of vision to give behavior to synthetic actors seems similar to the use of vision for intelligent mobile robots [30 31], it is quite different. This is because the vision of the synthetic actor is itself a synthetic vision. Using a synthetic vision allow us to skip all the problems of pattern recognition and distance detection, problems which still are the most difficult parts in robotics vision. However some interesting work has been done in the topic of intelligent mobile robots [32], especially for actionperception coordination problems. The concept of synthetic vision was first introduced by Renault et al. [33] as a main information channel between the environment and the virtual actor. More recently, several authors have adopted this approach for simulating the behavior of groups [34], fishes[35], dog [36]. In [33], each pixel of the vision input has the semantic information giving the object projected on this pixel, and numerical information giving the distance to this object. So, it is easy to know, for example, that there is a table just in front at 3 meters. The synthetic actor perceives his environment from a small window in which the environment is rendered from his point of view. As he can access z-buffer values of the pixels, the color of the pixels and his own position, he can locate visible objects in his 3D environment. Noser et al. [37] also propose the use of an octree as the internal representation of the environment seen by an actor because it offers several interesting features. The octree has to represent the visual memory of an actor in a 3D environment with static and dynamic objects. Objects in this environment can grow, shrink, move or disappear. To illustrate the capabilities of the synthetic vision system, the authors have developed several examples: the actor going out of a maze, walking on sparse foot locations and playing tennis Virtual Audition In real life, the behavior of people or animals is very often influenced by sounds. To recreate the synthetic audition, in a first step, we had to model a sound environment where the synthetic actor can directly access to positional and semantic sound source information of a audible sound event. Now, our virtual actors are able to hear [38]. The sound renderer takes into account the real time constraints. So it is capable to render each time increment for each microphone in "real time" by taking into account the final propagation speed of sound and the moving sound sources and microphones Virtual Tactile and Haptic Our aim is to build a behavioral model based on tactile sensory input received at the level of skin from the environment. This sensory information can be used in tasks as touching objects, pressing buttons or kicking objects. A very important, but special case is the contact between the hands and objects during the grasping process. Our approach [39] is based on sphere multi-sensors. These sensors are considered as a group of objects attached to the articulated figure. A sensor is activated for any collision with other objects or sensors. Each sphere sensor is fitted to its associated joint shape with different radii. This configuration is important in our method because when a sensor is activated in a finger, only the articulations above it stop moving, while others can still move. By doing this way, all the fingers are finally positioned naturally around the object. These multisensors have been integrated in a general Autonomous Virtual Humans

9 methodology for automatic grasping of objects by the synthetic actors. This methodology is based on the following steps: Based on virtual sensors, typically the vision, the actor decides which object to grasp. The actor may have to walk in order to go near the object Inverse kinematics is used to find the final arm posture Based on a grasp taxonomy [40], the system decides the way the actor grasps the object (see Figure 6). For example, it decides to use a pinch when the object is too small to be grasped by more than two fingers or to use two hands when the object is large. Using multi-sensors, the fingers are adjusted in order to have an accurate grasping and to give feedback to the actor. Fig. 6. Examples of grasping different objects Virtual vision is especially useful for navigation. The navigation can be local or global. With a local navigation, the actor goes straight on to his goal and it is possible that he cannot reach it. With a global navigation, the actor first tries to find a path to his goal and if the path exists, the actor follows it until he reaches the goal position or until he detects a collision by his vision. During global navigation the actor memorizes his perceived environment by voxelizing it, based on his virtual vision. When the actor evolves in his environment, a simple walking model is not sufficient, the actor has to adapt his trajectory based on the variations of terrain by bypassing, jumping or climbing the obstacles he meets. The bypassing of obstacles consists in changing the direction and velocity of the walking of the actor. Jumping and climbing correspond to more complex motion. These actions should generate parameterized motion depending on the height and the length of the obstacle for a jump and the height and location of the feet for climbing the obstacle. These characteristics are determined by the actor from his perception. Figure 7 shows an actress walking in a landscape. Fig. 7. Actress walking Autonomous Virtual Humans - 9 -

10 5 Interactive Perceptive Actors We define an interactive perceptive synthetic actor as an actor aware of other actors and real people. Such an actor is also assumed to be autonomous of course. Moreover, he is able to communicate interactively with the other actors whatever their type and the real people. 5.1 Intercommunication between Synthetic Actors Behaviors may be also dependent on the emotional state of the actor. We also developed a model of nonverbal communication [41]. The believability of virtual actors is improved by their capability to interpret and use a nonverbal language. A nonverbal communication is concerned with postures and their indications on what people are feeling. Postures are the means to communicate and are defined by a specific position of the arms and legs and angles of the body. Usually, people don t use consciously nonverbal communication, but they instinctively understand it to a considerable extent and respond to it without any explicit reasoning. These nonverbal communication is essential to drive the interaction between people without contact (Figure 8) or with contact (Figure 9). Fig. 8. Nonverbal intercommunication Fig. 9. Interaction between two actors 5.2 Sensing the Real World for a Synthetic Actor The real people are of course easily aware of the actions of the synthetic actors through VR tools like Head-mounted displays, but one major problem to solve is to make the virtual actors conscious of the behavior of the real people. Virtual actors should sense the participants through their virtual sensors. Autonomous and perceptive actors can have different degrees of autonomy and different sensing channels to the environment. Let us now consider the case of a participant playing tennis with an interactive perceptive actor Autonomous Virtual Humans

11 The participant can participate in VR by the head-mounted display and the earphones. He cannot get any internal VR information. His only source of knowledge from the VR is communicated by the vision, the sound, and some tactile sensory information. His behavior is strongly influenced by this sensory input and his proper intelligence. In order to process the vision of the perceptive actor in a similar way than the vision of the participant, we need to have a different model. In this case, the only information obtained by the perceptive actor will be through the virtual vision looking at the participant. Such a perceptive actor would be independent of each VR representation and he could in the same manner communicate with participants and other perceptive actors. Perceptive actors and participants may easily be exchanged as demonstrated with the interactive game facility. For virtual audition, we encounter the same problem as in virtual vision. The real time constraints in VR demand fast reaction to sound signals and fast recognition of the semantic it carries. Concerning the tactile sensor, we may consider the following example: the participant places an object into the Virtual Space using a CyberGlove and the autonomous virtual actor will try to grasp it and put it on a virtual table for example. The actor interacts with the environment by grasping the object and moving it. At the beginning of interactive grasping, only the hand center sensor is active. The six palm values from the CyberGlove are used to move it towards the object. Inverse kinematics update the arm postures from hand center movement. After the sensor is activated, the hand is close enough to the object final frame. The hand center sensor is deactivated and multi-sensors on hand are now used, to detect sensor object collision. As an example, we have produced a fighting between a real person and an autonomous actor. The motion of the real person is captured using a Flock of Birds. The gestures are recognized by the system and the information is transmitted to the virtual actor who is able to react to the gestures and decide which attitude to do. Figure 10 shows an example. Such an approach is based on the more general concept of real-time recognition-based metaphor [24], a method consisting of recording input data from a VR device in real-time. The input data are analyzed. Based on the meaning of the input data, a corresponding directive is executed. Fig. 10. Fight between a participant and an interactive perceptive actor 5.3 Real-Time Synthetic Actors in Distributed Virtual Environments The VLNET [27, 42] (Virtual Life NETwork) system supports a networked shared virtual environment that allows multiple users to interact with each other and their surrounding in real time. The users are represented by 3D virtual human actors, which serve as agents to interact with the environment and other agents. The agents have similar appearance and behaviors with the real humans, to support the sense of presence of the users in the environment. In addition to guided actors, the environment can also include autonomous and interactive perceptive actors used as a friendly user interface to different services. Virtual humans can also be used in order to represent the currently unavailable partners, allowing asynchronous cooperation between distant partners. Autonomous Virtual Humans

12 5.4 Playing Interactive Tennis with Marilyn As application, we selected a virtual tennis game with a guided actor, the interactive perceptive actress Marilyn, and an interactive perceptive referee, both with a virtual vision. Figure 11 shows an interactive session. The geometric characteristics of the tennis court make part of the players knowledge. For the dynamics simulation of the ball, gravity, net, ground and the racquet we use physics-based calculations. The tracking of the ball by the vision system is controlled by a special automata that tracks the ball, estimates the collision time and collision point of ball and racquet and performs successfully a hit with given force and a given resulting ball direction. The referee judges the game by following the ball with his vision system. He updates the state of the match when he "hears" a ball collision event (ball - ground, ball - net, ball - racket) according to what he sees and his knowledge of a tennis match, and he communicates his decisions and the state of the game by "spoken words" (sound events). Marilyn can also hear sound events and obeys the decisions of the referee. Fig. 11. Tennis game with real-time participant and interactive perceptive actors 6 The Future of Actors The ultimate objective in creating realistic and believable synthetic actors is to build intelligent autonomous virtual humans with adaptation, perception and memory. These actors should be able to act freely and emotionally. Ideally, they should be conscious and unpredictable. But, how far are we from such a ideal situation? Our interactive perceptive actors are able to perceive the virtual world, the people living in this world and in the real world. They may act based on their perception in an autonomous manner. Their intelligence is constrained and limited to the results obtained in the development of new methods of Artificial Intelligence. However, the representation under the form of virtual actors is a way of visually evaluating the progress. Intelligent actors are able to learn or understand very simple situations. Memory is generally defined as the power or process of reproducing or recalling what has been learned and retained especially through associative mechanisms. We have seen that emotional aspects may be important in nonverbal intercommunication. Emotions are also essential in facial animation. However, a real emotion should be considered as a state of feeling, a psychic and physical reaction subjectively experienced as strong feeling and physiologically involving changes that prepare the body for immediate vigorous action. In this case, we are far from implementing truly emotional actors. Finally, actors in the future should be adaptive, conscious and free. An actor is adaptive as long as it may "survive" in more or less unpredictable and dangerous environments. According to Alexander [43], a conscious actor should be aware especially of something within himself or characterized by sensation, emotion, volition, and thought. An actor may be considered as free if his future behaviour is unpredictable to somebody. Acknowledgments The author is grateful to the people who contributed to this work, in particular Pascal Bécheiraz, Ronan Boulic, Tolga Çapin, Zhyong Huang, Joël François, Shen Jianhua, Alexis Kappauf, Tom Molet, Serge Rezzonico, and Hansrudi Noser. The research was supported by the Swiss National Science Research Foundation, the Federal Office for Education and Science, and is part of the Esprit Project HUMANOID-2 and the ACTS project COVEN Autonomous Virtual Humans

13 References [1] N. Magnenat-Thalmann, D. Thalmann, P.Bergeron, DREAM FLIGHT: a Fictional Film Produced by 3D Computer Animation, Proc. Computer Graphics '82, London, 1983, pp [2] N. Magnenat-Thalmann, D. Thalmann, The Use of 3D High-Level Graphical Types in the MIRA Animation System, IEEE Computer Graphics and Applications, Vol. 3, No 9, 1983, pp [3] N. Magnenat-Thalmann, D. Thalmann, The Direction of Synthetic Actors in the Film Rendez-vous à Montréal, IEEE Computer Graphics and Applications, Vol.7, No 12, 1987, pp [4] R.Maiocchi R D Character Animation Using Motion Capture, in: Magnenat Thalmann, N. and Thalmann D. (eds) Interactive Computer Animation, Prentice Hall. [5] N:Magnenat Thalmann N., Thalmann D. (1995) Digital Actors for Interactive Television, Proc. IEEE, Special Issue on Digital Television, Part 2, July 1995, pp [6] M.Slater, M.Usoh, Body Centred Interaction in Immersive Virtual Environments, in: (Magnenat Thalmann and Thalmann, eds) Artificial Life and Virtual Reality, J.Wiley, Chichester, 1994, pp [7] A.LeBlanc., P.Kalra, N.Magnenat Thalmann, D.Thalmann (1991) Sculpting With the Ball & Mouse Metaphor, Proc. Graphics Interface '91, Calgary, Canada, pp [8] J.E.Chadwick., D.R.Haumann, R.E.Parent, Layered construction for deformable animated character, Computer Graphics, 23, (3), 1989, pp [9] M.P.Gascuel, A.Verroust, C.Puech, A modelling system for complex deformable bodies suited to animation and collision processing, Journal of Visualization and Computer Animation, Vol 2,1991, pp [10] S.Jianhua, D.Thalmann, Interactive Shape Design Using Metaballs and Splines, Proc. Eurographics Workshop on Implicit Surfaces 95, Grenoble, pp [11] D.Thalmann, S.Jianhua, E.Chauvineau, Fast Human Body Deformations for Animation and VR Applications, Proc. Computer Graphics International 96, IEEE Computer Society Press, June [12] N.I.Badler, J.D.Korein, J.U.Korein, G.M.Radack, L.S.Brotman (1985) Positioning and Animating Figures in a Task-oriented Environment, The Visual Computer, Vol.1, No.4, pp [13] W.W.Armstrong, M.Green, R.Lake (1987) Near real-time Control of Human Figure Models, IEEE CG&A, Vol.7, No6, pp [14] A.Witkin, M.Kass (1988) Spacetime Constraints, Proc.SIGGRAPH '88, pp [15] M.F.Cohen (1992) Interactive Spacetime Control for Animation, Proc. Siggraph'92, pp [16 P. Kalra, A. Mangili, N. Magnenat Thalmann, D. Thalmann, Simulation of Facial Muscle Actions Based on Rational Free Form Deformations, Proc. Eurographics '92, Cambridge, pp [17] Y.Wu, N.Magnenat Thalmann, D. Thalmann, A Plastic Visco-Elastic Model for Wrinkles in Facial Animation and Skin Aging, Journal of Visualization and Computer Animation, John Wiley, Vol.6, No4, 1995, pp [18] P.Volino, M.Courchesne, N.Magnenat Thalmann, Versatile and Efficient Techniques for Simulating Cloth and Other Deformable Objects, Computer Graphics, 29, 1995, pp [19] P.Volino, N.Magnenat Thalmann, S.Jianhua, D.Thalmann, A 3D Evolutive System for Simulating Deformable Clothes on Virtual Actors, IEEE Computer Graphics & Applications, 1996 (to appear) [20] N.Magnenat Thalmann, D.Thalmann, Virtual Actors Living in a Real World, Proc. Computer Animation 95, IEEE Computer Society Press, April 1995, pp [21] P.Maes, T.Darrell, B.Blumberg, A.Pentland, The ALIVE system: Full-Body Interaction with Autonomous Agents, Proc. Computer Animation 95, IEEE Computer Society Press, 1995, pp [22] D.Zeltzer (1985) Towards an Integrated View of 3D Computer Animation, The Visual Computer, Vol. 1, No4, pp [23] N.Magnenat Thalmann, D.Thalmann, Complex Models for Visualizing Synthetic Actors. In IEEE Computer Graphics and Applications, Vol.11, No5, 1991, pp [24] D.Thalmann, Using Virtual Reality Techniques in the Animation Process in: Virtual Reality Systems (Earnshaw R, Gigante M, Jones H eds), Academic Press, pp [25] I.Pandzic, T.Çapin, N.Magnenat Thalmann, D. Thalmann, VLNET: A Networked Multimedia 3D Environment with Virtual Humans, Proc. Multi-Media Modeling MMM 95, World Scientific, Singapore, 1995, pp [26] R.Boulic, N.Magnenat-Thalmann, D.Thalmann, A Global Human Walking Model with Real Time Kinematic Personification, The Visual Computer, Vol.6(6),1990, pp [27] M.Slater, A.Steed, M.Usoh, The Virtual Treadmill: A Naturalistic Method for Navigation in Immersive Virtual Environments, in: M.Göbel (ed.) Virtual Environments 95, Springer, Wien, 1995, pp [28] C.Reynolds, Flocks, Herds, and Schools: A Distributed Behavioral Model, Proc.SIGGRAPH '87, Computer Graphics, Vol.21, No4, 1987, pp [29] D. Thalmann, Virtual Sensors: A Key Tool for the Artificial Life of Virtual Actors, Proc. Pacific Graphics 95, Seoul, Korea, 1995, pp [30] I.Horswill, A Simple, Cheap, and Robust Visual Navigation System, in: From Animals to Animats 2, Proc. 2nd Intern. Conf. on Simulation of Adaptive Behavior, MIT Press, 1993, pp [31] S.Tsuji, S.Li, Memorizing and Representing Route Scenes, in:meyer J.A.et al.(eds) From Animals to Animats, Proc.2nd Intern.Conf.on Simulation of Adaptive Behavior, MIT Press, 1993, pp [32] J.L.Crowley, Navigation for an Intelligent Mobile Robot, IEEE journal of Robotics and Automation, Vol. RA-1, No. 1, 1987, pp Autonomous Virtual Humans

14 [33] O.Renault, N.Magnenat Thalmann, D.Thalmann, A Vision-based Approach to Behavioural Animation, The Journal of Visualization and Computer Animation, Vol 1, No 1, 1990, pp [34] Reynolds C.W. (1993) An Evolved, Vision-Based Behavioral Model of Coordinated Group Motion, in: Meyer J.A. et al. (eds) From Animals to Animats, Proc. 2nd International Conf. on Simulation of Adaptive Behavior, MIT Press, pp [35] Tu X., Terzopoulos D., Artificial Fishes: Physics, Locomotion, Perception, Behavior, Proc. SIGGRAPH '94, Computer Graphics, 1994, pp [36] B.M.Blumberg, T.A,Galyean, Multi-Level Direction of Autonomous Creatures for Real-Time Virtual Environments, Proc. SIGGRAPH 95, 1995, pp [37] H.Noser, O.Renault, D.Thalmann, N.Magnenat Thalmann, Navigation for Digital Actors based on Synthetic Vision, Memory and Learning, Computers and Graphics, Vol.19, No1, 1995, pp [38] H.Noser, D.Thalmann, Synthetic Vision and Audition for Digital Actors, Proc.Eurographics 95, 1995, pp [39] Z.Huang, R.Boulic, N.Magnenat Thalmann, D.Thalmann, A Multi-sensor Approach for Grasping and 3D Interaction, Proc. CGI 95, Academic Press, 1995, pp [40] R.Mas., D.Thalmann, A Hand Control and Automatic Grasping System for Synthetic Actors, Proc.Eurographic'94, 1994, pp [41] P.Bécheiraz, D.Thalmann, A Model of Nonverbal Communication and Interpersonal Relationship between Virtual Actors, Proc.Computer Animation 96, IEEE Computer Society Press, [42] D.Thalmann, T.Capin., N.Magnenat-Thalmann., I.S.Pandzic, Participant, User-guided, and Autonomous Actors in the Virtual Life Network VLNET, Proc.ICAT/VRST 95, Chiba, Japan, 1995, pp [43] I.Aleksander, Artificial Consciousness, in: (Magnenat Thalmann and Thalmann, eds) Artificial Life and Virtual Reality, John Wiley, Chichester, 1994, pp Autonomous Virtual Humans

"Towards Intelligent Virtual Actors"

Towards Intelligent Virtual Actors "Towards Intelligent Virtual Actors" Daniel THALMANN Computer Graphics Lab, EPFL CH 1015 Lausanne Switzerland thalmann@lig.di.epfl.ch In this paper, we try to define the concept of virtual actors and show

More information

Towards Autonomous, Perceptive, and Intelligent Virtual Actors

Towards Autonomous, Perceptive, and Intelligent Virtual Actors 1 Towards Autonomous, Perceptive, and Intelligent Virtual Actors Daniel Thalmann Computer Graphics Lab, EPFL - LIG Lausanne, Switzerland thalmann@lig.di.epfl.ch Hansrudi Noser Multimedia Laboratorium University

More information

An Interactive Interface for Directing Virtual Humans

An Interactive Interface for Directing Virtual Humans An Interactive Interface for Directing Virtual Humans Gael Sannier 1, Selim Balcisoy 2, Nadia Magnenat-Thalmann 1, Daniel Thalmann 2 1) MIRALab, University of Geneva, 24 rue du Général Dufour CH 1211 Geneva,

More information

The Simulation of a Virtual TV Presentor

The Simulation of a Virtual TV Presentor MIRALab Copyright Information 1998 The Simulation of a Virtual TV Presentor Abstract Nadia Magnenat Thalmann, Prem Kalra MIRALab, University of Geneva This paper presents the making of six short sequences

More information

CONSTRUCTING VIRTUAL HUMAN LIFE SIMULATIONS

CONSTRUCTING VIRTUAL HUMAN LIFE SIMULATIONS CONSTRUCTING VIRTUAL HUMAN LIFE SIMULATIONS Marcelo Kallmann, Etienne de Sevin and Daniel Thalmann Swiss Federal Institute of Technology (EPFL), Computer Graphics Lab (LIG), Lausanne, Switzerland, CH-1015,

More information

Modeling the Virtual World

Modeling the Virtual World Modeling the Virtual World Joaquim Madeira November, 2013 RVA - 2013/2014 1 A VR system architecture Modeling the Virtual World Geometry Physics Haptics VR Toolkits RVA - 2013/2014 2 VR object modeling

More information

EPFL. When an animator species an animation sequence, it privileges when animating the Virtual. body, the privileged information to be manipulated

EPFL. When an animator species an animation sequence, it privileges when animating the Virtual. body, the privileged information to be manipulated Realistic Human Body Modeling P. Fua, R. Plankers, and D. Thalmann Computer Graphics Lab (LIG) EPFL CH-1015 Lausanne, Switzerland 1 Introduction Synthetic modeling of human bodies and the simulation of

More information

The Virtual Human as a Multimodal Interface

The Virtual Human as a Multimodal Interface The Virtual Human as a Multimodal Interface Daniel Thalmann Computer Graphics Lab, EPFL Lausanne, Switzerland thalmann @ li9.di.epfl.ch http://ligwww.epfl.ch ABSTRACT This paper discusses the main issues

More information

Design and Implementation of Synthetic Humans for Virtual Environments and Simulation Systems

Design and Implementation of Synthetic Humans for Virtual Environments and Simulation Systems Design and Implementation of Synthetic Humans for Virtual Environments and Simulation Systems SPYROS VOSINAKIS, THEMIS PANAYIOTOPOULOS Department of Informatics University of Piraeus 80 Karaoli & Dimitriou

More information

The HUMANOID Environment for Interactive Animation of Multiple Deformable Human Characters

The HUMANOID Environment for Interactive Animation of Multiple Deformable Human Characters The HUMANOID Environment for Interactive Animation of Multiple Deformable Human Characters R. Boulic 1, T. Capin 1, Z. Huang 1, P. Kalra 2, B. Lintermann 3, N. Magnenat-Thalmann 2, L. Moccozet 2, T. Molet

More information

Animation, Motion Capture, & Inverse Kinematics. Announcements: Quiz

Animation, Motion Capture, & Inverse Kinematics. Announcements: Quiz Animation, Motion Capture, & Inverse Kinematics Announcements: Quiz On Tuesday (3/10), in class One 8.5x11 sheet of notes allowed Sample quiz (from a previous year) on website Focus on reading comprehension

More information

A tool for constructing 3D Environments with Virtual Agents

A tool for constructing 3D Environments with Virtual Agents A tool for constructing 3D Environments with Virtual Agents Spyros Vosinakis, Themis Panayiotopoulos Knowledge Engineering Laboratory, Department of Informatics, University of Piraeus, 80 Karaoli & Dimitriou

More information

Computer Graphics. Si Lu. Fall uter_graphics.htm 11/27/2017

Computer Graphics. Si Lu. Fall uter_graphics.htm 11/27/2017 Computer Graphics Si Lu Fall 2017 http://web.cecs.pdx.edu/~lusi/cs447/cs447_547_comp uter_graphics.htm 11/27/2017 Last time o Ray tracing 2 Today o Animation o Final Exam: 14:00-15:30, Novermber 29, 2017

More information

Announcements: Quiz. Animation, Motion Capture, & Inverse Kinematics. Last Time? Today: How do we Animate? Keyframing. Procedural Animation

Announcements: Quiz. Animation, Motion Capture, & Inverse Kinematics. Last Time? Today: How do we Animate? Keyframing. Procedural Animation Announcements: Quiz Animation, Motion Capture, & Inverse Kinematics On Friday (3/1), in class One 8.5x11 sheet of notes allowed Sample quiz (from a previous year) on website Focus on reading comprehension

More information

Homework 2 Questions? Animation, Motion Capture, & Inverse Kinematics. Velocity Interpolation. Handing Free Surface with MAC

Homework 2 Questions? Animation, Motion Capture, & Inverse Kinematics. Velocity Interpolation. Handing Free Surface with MAC Homework 2 Questions? Animation, Motion Capture, & Inverse Kinematics Velocity Interpolation Original image from Foster & Metaxas, 1996 In 2D: For each axis, find the 4 closest face velocity samples: Self-intersecting

More information

Last Time? Animation, Motion Capture, & Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

Last Time? Animation, Motion Capture, & Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation Last Time? Animation, Motion Capture, & Inverse Kinematics Navier-Stokes Equations Conservation of Momentum & Mass Incompressible Flow Today How do we animate? Keyframing Procedural Animation Physically-Based

More information

SYNTHETIC VISION AND EMOTION CALCULATION IN INTELLIGENT VIRTUAL HUMAN MODELING

SYNTHETIC VISION AND EMOTION CALCULATION IN INTELLIGENT VIRTUAL HUMAN MODELING SYNTHETIC VISION AND EMOTION CALCULATION IN INTELLIGENT VIRTUAL HUMAN MODELING Y. Zhao, J. Kang and D. K. Wright School of Engineering and Design, Brunel University, Uxbridge, Middlesex, UB8 3PH, UK ABSTRACT

More information

THE PROBLEMATICS OF HUMAN PROTOTYPING AND ANIMATION

THE PROBLEMATICS OF HUMAN PROTOTYPING AND ANIMATION THE PROBLEMATICS OF HUMAN PROTOTYPING AND ANIMATION Nadia Magnenat-Thalmann Centre Universitaire d'informatique Université de Genève, Suisse Daniel Thalmann Laboratoire d'infographie Ecole Polytechnique

More information

Last Time? Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

Last Time? Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation Last Time? Inverse Kinematics Navier-Stokes Equations Conservation of Momentum & Mass Incompressible Flow Today How do we animate? Keyframing Procedural Animation Physically-Based Animation Forward and

More information

To Do. History of Computer Animation. These Lectures. 2D and 3D Animation. Computer Animation. Foundations of Computer Graphics (Spring 2010)

To Do. History of Computer Animation. These Lectures. 2D and 3D Animation. Computer Animation. Foundations of Computer Graphics (Spring 2010) Foundations of Computer Graphics (Spring 2010) CS 184, Lecture 24: Animation http://inst.eecs.berkeley.edu/~cs184 To Do Submit HW 4 (today) Start working on HW 5 (can be simple add-on) Many slides courtesy

More information

Motion Capture & Simulation

Motion Capture & Simulation Motion Capture & Simulation Motion Capture Character Reconstructions Joint Angles Need 3 points to compute a rigid body coordinate frame 1 st point gives 3D translation, 2 nd point gives 2 angles, 3 rd

More information

A Comparison of Design Strategies for 3D Human Motions*

A Comparison of Design Strategies for 3D Human Motions* A Comparison of Design Strategies for 3D Human Motions* Ronan Boulic, Zhiyong Huang, Daniel Thalmann Computer Graphics Lab (LIG), Swiss Federal Institute of Technology (EPFL) CH-1015 Lausanne, Switzerland

More information

Last Time? Animation, Motion Capture, & Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation

Last Time? Animation, Motion Capture, & Inverse Kinematics. Today. Keyframing. Physically-Based Animation. Procedural Animation Last Time? Animation, Motion Capture, & Inverse Kinematics Navier-Stokes Equations Conservation of Momentum & Mass Incompressible Flow Today How do we animate? Keyframing Procedural Animation Physically-Based

More information

A Behavioral Interface to Simulate Agent-Object Interactions in Real Time

A Behavioral Interface to Simulate Agent-Object Interactions in Real Time A Behavioral Interface to Simulate Agent-Object Interactions in Real Time Marcelo Kallmann and Daniel Thalmann EPFL Computer Graphics Lab LIG CH-1015 Lausanne Switzerland {kallmann, thalmann}@lig.di.epfl.ch

More information

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation

Human body animation. Computer Animation. Human Body Animation. Skeletal Animation Computer Animation Aitor Rovira March 2010 Human body animation Based on slides by Marco Gillies Human Body Animation Skeletal Animation Skeletal Animation (FK, IK) Motion Capture Motion Editing (retargeting,

More information

Data-driven Approaches to Simulation (Motion Capture)

Data-driven Approaches to Simulation (Motion Capture) 1 Data-driven Approaches to Simulation (Motion Capture) Ting-Chun Sun tingchun.sun@usc.edu Preface The lecture slides [1] are made by Jessica Hodgins [2], who is a professor in Computer Science Department

More information

A Memory Model for Autonomous Virtual Humans

A Memory Model for Autonomous Virtual Humans A Memory Model for Autonomous Virtual Humans Christopher Peters Carol O Sullivan Image Synthesis Group, Trinity College, Dublin 2, Republic of Ireland email: {christopher.peters, carol.osullivan}@cs.tcd.ie

More information

Character Animation 1

Character Animation 1 Character Animation 1 Overview Animation is a big topic We will concentrate on character animation as is used in many games today humans, animals, monsters, robots, etc. Character Representation A character

More information

Character Animation. Presented by: Pam Chow

Character Animation. Presented by: Pam Chow Character Animation Presented by: Pam Chow Overview Animation is a big topic We will concentrate on character animation as is used in many games today humans, animals, monsters, robots, etc. PLAZMO AND

More information

The 3D rendering pipeline (our version for this class)

The 3D rendering pipeline (our version for this class) The 3D rendering pipeline (our version for this class) 3D models in model coordinates 3D models in world coordinates 2D Polygons in camera coordinates Pixels in image coordinates Scene graph Camera Rasterization

More information

CS770/870 Spring 2017 Animation Basics

CS770/870 Spring 2017 Animation Basics Preview CS770/870 Spring 2017 Animation Basics Related material Angel 6e: 1.1.3, 8.6 Thalman, N and D. Thalman, Computer Animation, Encyclopedia of Computer Science, CRC Press. Lasseter, J. Principles

More information

CS770/870 Spring 2017 Animation Basics

CS770/870 Spring 2017 Animation Basics CS770/870 Spring 2017 Animation Basics Related material Angel 6e: 1.1.3, 8.6 Thalman, N and D. Thalman, Computer Animation, Encyclopedia of Computer Science, CRC Press. Lasseter, J. Principles of traditional

More information

animation projects in digital art animation 2009 fabio pellacini 1

animation projects in digital art animation 2009 fabio pellacini 1 animation projects in digital art animation 2009 fabio pellacini 1 animation shape specification as a function of time projects in digital art animation 2009 fabio pellacini 2 how animation works? flip

More information

Animation. CS 465 Lecture 22

Animation. CS 465 Lecture 22 Animation CS 465 Lecture 22 Animation Industry production process leading up to animation What animation is How animation works (very generally) Artistic process of animation Further topics in how it works

More information

1. ABSTRACT 3. CROWD MODEL

1. ABSTRACT 3. CROWD MODEL Crowd Modelling in Collaborative Virtual Environments Soraia R. Musse Christian Babski Tolga Çapin Daniel Thalmann {soraia,babski,capin,thalmann}@lig.di.epfl.ch Computer Graphics Lab. Swiss Federal Institute

More information

BONE CONTROLLER ASSET VERSION 0.1 REV 1

BONE CONTROLLER ASSET VERSION 0.1 REV 1 Foreword Thank you for purchasing the Bone Controller! I m an independent developer and your feedback and support really means a lot to me. Please don t ever hesitate to contact me if you have a question,

More information

Animation. Representation of objects as they vary over time. Traditionally, based on individual drawing or photographing the frames in a sequence

Animation. Representation of objects as they vary over time. Traditionally, based on individual drawing or photographing the frames in a sequence 6 Animation Animation Representation of objects as they vary over time Traditionally, based on individual drawing or photographing the frames in a sequence Computer animation also results in a sequence

More information

COMP 175 COMPUTER GRAPHICS. Lecture 10: Animation. COMP 175: Computer Graphics March 12, Erik Anderson 08 Animation

COMP 175 COMPUTER GRAPHICS. Lecture 10: Animation. COMP 175: Computer Graphics March 12, Erik Anderson 08 Animation Lecture 10: Animation COMP 175: Computer Graphics March 12, 2018 1/37 Recap on Camera and the GL Matrix Stack } Go over the GL Matrix Stack 2/37 Topics in Animation } Physics (dynamics, simulation, mechanics)

More information

Overview. Agents and Avatars. Why IVAs? IVAS. Agents and Avatars Believability v naturalism Building the body H-anim Moving and sensing.

Overview. Agents and Avatars. Why IVAs? IVAS. Agents and Avatars Believability v naturalism Building the body H-anim Moving and sensing. Overview Agents and Avatars Ruth Aylett Agents and Avatars Believability v naturalism Building the body H-anim Moving and sensing IVAS Why IVAs? Intelligent Virtual Agents (IVAs) Also: Synthetic characters

More information

Animation. CS 4620 Lecture 33. Cornell CS4620 Fall Kavita Bala

Animation. CS 4620 Lecture 33. Cornell CS4620 Fall Kavita Bala Animation CS 4620 Lecture 33 Cornell CS4620 Fall 2015 1 Announcements Grading A5 (and A6) on Monday after TG 4621: one-on-one sessions with TA this Friday w/ prior instructor Steve Marschner 2 Quaternions

More information

7 Modelling and Animating Human Figures. Chapter 7. Modelling and Animating Human Figures. Department of Computer Science and Engineering 7-1

7 Modelling and Animating Human Figures. Chapter 7. Modelling and Animating Human Figures. Department of Computer Science and Engineering 7-1 Modelling and Animating Human Figures 7-1 Introduction Modeling and animating an articulated figure is one of the most formidable tasks that an animator can be faced with. It is especially challenging

More information

Applications. Systems. Motion capture pipeline. Biomechanical analysis. Graphics research

Applications. Systems. Motion capture pipeline. Biomechanical analysis. Graphics research Motion capture Applications Systems Motion capture pipeline Biomechanical analysis Graphics research Applications Computer animation Biomechanics Robotics Cinema Video games Anthropology What is captured?

More information

The Application Research of 3D Simulation Modeling Technology in the Sports Teaching YANG Jun-wa 1, a

The Application Research of 3D Simulation Modeling Technology in the Sports Teaching YANG Jun-wa 1, a 4th National Conference on Electrical, Electronics and Computer Engineering (NCEECE 2015) The Application Research of 3D Simulation Modeling Technology in the Sports Teaching YANG Jun-wa 1, a 1 Zhengde

More information

Real-time Talking Head Driven by Voice and its Application to Communication and Entertainment

Real-time Talking Head Driven by Voice and its Application to Communication and Entertainment ISCA Archive Real-time Talking Head Driven by Voice and its Application to Communication and Entertainment Shigeo MORISHIMA Seikei University ABSTRACT Recently computer can make cyberspace to walk through

More information

Multimedia Technology CHAPTER 4. Video and Animation

Multimedia Technology CHAPTER 4. Video and Animation CHAPTER 4 Video and Animation - Both video and animation give us a sense of motion. They exploit some properties of human eye s ability of viewing pictures. - Motion video is the element of multimedia

More information

Cloth Simulation. Tanja Munz. Master of Science Computer Animation and Visual Effects. CGI Techniques Report

Cloth Simulation. Tanja Munz. Master of Science Computer Animation and Visual Effects. CGI Techniques Report Cloth Simulation CGI Techniques Report Tanja Munz Master of Science Computer Animation and Visual Effects 21st November, 2014 Abstract Cloth simulation is a wide and popular area of research. First papers

More information

M I RA Lab. Speech Animation. Where do we stand today? Speech Animation : Hierarchy. What are the technologies?

M I RA Lab. Speech Animation. Where do we stand today? Speech Animation : Hierarchy. What are the technologies? MIRALab Where Research means Creativity Where do we stand today? M I RA Lab Nadia Magnenat-Thalmann MIRALab, University of Geneva thalmann@miralab.unige.ch Video Input (face) Audio Input (speech) FAP Extraction

More information

MODELING AND HIERARCHY

MODELING AND HIERARCHY MODELING AND HIERARCHY Introduction Models are abstractions of the world both of the real world in which we live and of virtual worlds that we create with computers. We are all familiar with mathematical

More information

Computer Animation. Algorithms and Techniques. z< MORGAN KAUFMANN PUBLISHERS. Rick Parent Ohio State University AN IMPRINT OF ELSEVIER SCIENCE

Computer Animation. Algorithms and Techniques. z< MORGAN KAUFMANN PUBLISHERS. Rick Parent Ohio State University AN IMPRINT OF ELSEVIER SCIENCE Computer Animation Algorithms and Techniques Rick Parent Ohio State University z< MORGAN KAUFMANN PUBLISHERS AN IMPRINT OF ELSEVIER SCIENCE AMSTERDAM BOSTON LONDON NEW YORK OXFORD PARIS SAN DIEGO SAN FRANCISCO

More information

Virtual Interaction System Based on Optical Capture

Virtual Interaction System Based on Optical Capture Sensors & Transducers 203 by IFSA http://www.sensorsportal.com Virtual Interaction System Based on Optical Capture Peng CHEN, 2 Xiaoyang ZHOU, 3 Jianguang LI, Peijun WANG School of Mechanical Engineering,

More information

Facial Deformations for MPEG-4

Facial Deformations for MPEG-4 Facial Deformations for MPEG-4 Marc Escher, Igor Pandzic, Nadia Magnenat Thalmann MIRALab - CUI University of Geneva 24 rue du Général-Dufour CH1211 Geneva 4, Switzerland {Marc.Escher, Igor.Pandzic, Nadia.Thalmann}@cui.unige.ch

More information

Model-Based Human Motion Capture from Monocular Video Sequences

Model-Based Human Motion Capture from Monocular Video Sequences Model-Based Human Motion Capture from Monocular Video Sequences Jihun Park 1, Sangho Park 2, and J.K. Aggarwal 2 1 Department of Computer Engineering Hongik University Seoul, Korea jhpark@hongik.ac.kr

More information

AMSTERDAM BOSTON HEIDELBERG LONDON NEW YORK OXFORD PARIS SAN DIEGO SAN FRANCISCO SINGAPORE SYDNEY TOKYO F ^ k.^

AMSTERDAM BOSTON HEIDELBERG LONDON NEW YORK OXFORD PARIS SAN DIEGO SAN FRANCISCO SINGAPORE SYDNEY TOKYO F ^ k.^ Computer a jap Animation Algorithms and Techniques Second Edition Rick Parent Ohio State University AMSTERDAM BOSTON HEIDELBERG LONDON NEW YORK OXFORD PARIS SAN DIEGO SAN FRANCISCO SINGAPORE SYDNEY TOKYO

More information

Human Character Animation in 3D-Graphics: The EMOTE System as a Plug-in for Maya

Human Character Animation in 3D-Graphics: The EMOTE System as a Plug-in for Maya Hartmann - 1 Bjoern Hartman Advisor: Dr. Norm Badler Applied Senior Design Project - Final Report Human Character Animation in 3D-Graphics: The EMOTE System as a Plug-in for Maya Introduction Realistic

More information

COMPUTER ANIMATION 3 KEYFRAME ANIMATION, RIGGING, SKINNING AND CHARACTER ANIMATION. Rémi Ronfard, Animation, M2R MOSIG

COMPUTER ANIMATION 3 KEYFRAME ANIMATION, RIGGING, SKINNING AND CHARACTER ANIMATION. Rémi Ronfard, Animation, M2R MOSIG COMPUTER ANIMATION 3 KEYFRAME ANIMATION, RIGGING, SKINNING AND CHARACTER ANIMATION Rémi Ronfard, Animation, M2R MOSIG 2 Outline Principles of animation Keyframe interpolation Rigging, skinning and walking

More information

Motion Control with Strokes

Motion Control with Strokes Motion Control with Strokes Masaki Oshita Kyushu Institute of Technology oshita@ces.kyutech.ac.jp Figure 1: Examples of stroke-based motion control. Input strokes (above) and generated motions (below).

More information

Realistic Rendering and Animation of a Multi-Layered Human Body Model

Realistic Rendering and Animation of a Multi-Layered Human Body Model Realistic Rendering and Animation of a Multi-Layered Human Body Model Mehmet Şahin Yeşil and Uǧur Güdükbay Dept. of Computer Engineering, Bilkent University, Bilkent 06800 Ankara, Turkey email: syesil@alumni.bilkent.edu.tr,

More information

Animation COM3404. Richard Everson. School of Engineering, Computer Science and Mathematics University of Exeter

Animation COM3404. Richard Everson. School of Engineering, Computer Science and Mathematics University of Exeter Animation COM3404 Richard Everson School of Engineering, Computer Science and Mathematics University of Exeter R.M.Everson@exeter.ac.uk http://www.secamlocal.ex.ac.uk/studyres/com304 Richard Everson Animation

More information

Projection simulator to support design development of spherical immersive display

Projection simulator to support design development of spherical immersive display Projection simulator to support design development of spherical immersive display Wataru Hashimoto, Yasuharu Mizutani, and Satoshi Nishiguchi Faculty of Information Sciences and Technology, Osaka Institute

More information

Animation, Motion Capture, & Inverse Kinematics

Animation, Motion Capture, & Inverse Kinematics Animation, Motion Capture, & Inverse Kinematics Pop Worksheet! Teams of 2. SOMEONE YOU HAVEN T ALREADY WORKED WITH Enumerate all cases (including rotations) of the 2D version of Marching Cubes, labeling

More information

Computer Animation and Visualisation. Lecture 3. Motion capture and physically-based animation of characters

Computer Animation and Visualisation. Lecture 3. Motion capture and physically-based animation of characters Computer Animation and Visualisation Lecture 3. Motion capture and physically-based animation of characters Character Animation There are three methods Create them manually Use real human / animal motions

More information

Using Synthetic Vision for Autonomous Non-Player Characters in Computer Games

Using Synthetic Vision for Autonomous Non-Player Characters in Computer Games Using Synthetic Vision for Autonomous Non-Player Characters in Computer Games Sebastian Enrique 1, Alan Watt 2, Steve Maddock 2, Fabio Policarpo 3 1 Departamento de Computación, Facultad de Ciencias Exactas

More information

Multilevel Deformation Model Applied to Hand Simulation for Virtual Actors

Multilevel Deformation Model Applied to Hand Simulation for Virtual Actors MIRALab Copyright Information 1998 Multilevel Deformation Model Applied to Hand Simulation for Virtual Actors Laurent Moccozet, Nadia Magnenat Thalmann MIRALab, University of Geneva Abstract Until now,

More information

An Introduction to animation and motion blending

An Introduction to animation and motion blending An Introduction to animation and motion blending Stuart Bryson - 98082365 Information Technology Research Preparation University Technology, Sydney 14/06/2005 An introduction to animation and motion blending

More information

Computer Graphics Introduction. Taku Komura

Computer Graphics Introduction. Taku Komura Computer Graphics Introduction Taku Komura What s this course all about? We will cover Graphics programming and algorithms Graphics data structures Applied geometry, modeling and rendering Not covering

More information

Course Review. Computer Animation and Visualisation. Taku Komura

Course Review. Computer Animation and Visualisation. Taku Komura Course Review Computer Animation and Visualisation Taku Komura Characters include Human models Virtual characters Animal models Representation of postures The body has a hierarchical structure Many types

More information

animation computer graphics animation 2009 fabio pellacini 1 animation shape specification as a function of time

animation computer graphics animation 2009 fabio pellacini 1 animation shape specification as a function of time animation computer graphics animation 2009 fabio pellacini 1 animation shape specification as a function of time computer graphics animation 2009 fabio pellacini 2 animation representation many ways to

More information

animation computer graphics animation 2009 fabio pellacini 1

animation computer graphics animation 2009 fabio pellacini 1 animation computer graphics animation 2009 fabio pellacini 1 animation shape specification as a function of time computer graphics animation 2009 fabio pellacini 2 animation representation many ways to

More information

Animation Movie under Autodesk Maya

Animation Movie under Autodesk Maya Animation Movie under Autodesk Maya Julio Manuel Vega Pérez University Rey Juan Carlos, Móstoles (Madrid), Spain March 5, 2009 1 Abstract Computer graphics matured over many years and played an important

More information

Chapter 8 Visualization and Optimization

Chapter 8 Visualization and Optimization Chapter 8 Visualization and Optimization Recommended reference books: [1] Edited by R. S. Gallagher: Computer Visualization, Graphics Techniques for Scientific and Engineering Analysis by CRC, 1994 [2]

More information

Research Subject. Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group)

Research Subject. Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group) Research Subject Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group) (1) Goal and summary Introduction Humanoid has less actuators than its movable degrees of freedom (DOF) which

More information

Animation. Itinerary. What is Animation? What is Animation? Animation Methods. Modeling vs. Animation Computer Graphics Lecture 22

Animation. Itinerary. What is Animation? What is Animation? Animation Methods. Modeling vs. Animation Computer Graphics Lecture 22 15-462 Computer Graphics Lecture 22 Animation April 22, 2003 M. Ian Graham Carnegie Mellon University What is Animation? Making things move What is Animation? Consider a model with n parameters Polygon

More information

CSE452 Computer Graphics

CSE452 Computer Graphics CSE452 Computer Graphics Lecture 19: From Morphing To Animation Capturing and Animating Skin Deformation in Human Motion, Park and Hodgins, SIGGRAPH 2006 CSE452 Lecture 19: From Morphing to Animation 1

More information

Overview. Animation is a big topic We will concentrate on character animation as is used in many games today. humans, animals, monsters, robots, etc.

Overview. Animation is a big topic We will concentrate on character animation as is used in many games today. humans, animals, monsters, robots, etc. ANIMATION Overview Animation is a big topic We will concentrate on character animation as is used in many games today humans, animals, monsters, robots, etc. Character Representation A character is represented

More information

Kinematics & Motion Capture

Kinematics & Motion Capture Lecture 27: Kinematics & Motion Capture Computer Graphics and Imaging UC Berkeley CS184/284A, Spring 2017 Forward Kinematics (Slides with James O Brien) Forward Kinematics Articulated skeleton Topology

More information

Modeling of Humanoid Systems Using Deductive Approach

Modeling of Humanoid Systems Using Deductive Approach INFOTEH-JAHORINA Vol. 12, March 2013. Modeling of Humanoid Systems Using Deductive Approach Miloš D Jovanović Robotics laboratory Mihailo Pupin Institute Belgrade, Serbia milos.jovanovic@pupin.rs Veljko

More information

Animation. Itinerary Computer Graphics Lecture 22

Animation. Itinerary Computer Graphics Lecture 22 15-462 Computer Graphics Lecture 22 Animation April 22, 2003 M. Ian Graham Carnegie Mellon University Itinerary Review Basic Animation Keyed Animation Motion Capture Physically-Based Animation Behavioral

More information

Automatic visual recognition for metro surveillance

Automatic visual recognition for metro surveillance Automatic visual recognition for metro surveillance F. Cupillard, M. Thonnat, F. Brémond Orion Research Group, INRIA, Sophia Antipolis, France Abstract We propose in this paper an approach for recognizing

More information

CS 231. Deformation simulation (and faces)

CS 231. Deformation simulation (and faces) CS 231 Deformation simulation (and faces) Deformation BODY Simulation Discretization Spring-mass models difficult to model continuum properties Simple & fast to implement and understand Finite Element

More information

A PLASTIC-VISCO-ELASTIC MODEL FOR WRINKLES IN FACIAL ANIMATION AND SKIN AGING

A PLASTIC-VISCO-ELASTIC MODEL FOR WRINKLES IN FACIAL ANIMATION AND SKIN AGING MIRALab Copyright Information 1998 A PLASTIC-VISCO-ELASTIC MODEL FOR WRINKLES IN FACIAL ANIMATION AND SKIN AGING YIN WU, NADIA MAGNENAT THALMANN MIRALab, CUI, University of Geneva DANIEL THALMAN Computer

More information

Sketch-based Interface for Crowd Animation

Sketch-based Interface for Crowd Animation Sketch-based Interface for Crowd Animation Masaki Oshita 1, Yusuke Ogiwara 1 1 Kyushu Institute of Technology 680-4 Kawazu, Iizuka, Fukuoka, 820-8502, Japan oshita@ces.kyutech.ac.p ogiwara@cg.ces.kyutech.ac.p

More information

Embedding a Motion-Capture Interface in a Control Structure for Human Like Agent Behavior Achievement

Embedding a Motion-Capture Interface in a Control Structure for Human Like Agent Behavior Achievement Embedding a Motion-Capture Interface in a Control Structure for Human Like Agent Behavior Achievement LUIZ GARCIA 1 FERNANDO DA SILVA 2 3 ANTONIO OLIVEIRA 2 LUIZ VELHO 3 JONAS GOMES 3 1 Laboratoire d Analyse

More information

COM337 COMPUTER GRAPHICS Other Topics

COM337 COMPUTER GRAPHICS Other Topics COM337 COMPUTER GRAPHICS Other Topics Animation, Surface Details, Global Illumination Kurtuluş Küllü based on the book by Hearn, Baker, and Carithers There are some other important issues and topics that

More information

Virtual Marionettes: A System and Paradigm for Real-Time 3D Animation

Virtual Marionettes: A System and Paradigm for Real-Time 3D Animation Virtual Marionettes: A System and Paradigm for Real-Time 3D Animation Adi Bar-Lev, Alfred M. Bruckstein, Gershon Elber Computer Science Department Technion, I.I.T. 32000 Haifa, Israel Abstract This paper

More information

A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction

A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction Jaemin Lee and Ergun Akleman Visualization Sciences Program Texas A&M University Abstract In this paper we present a practical

More information

Crowd simulation. Taku Komura

Crowd simulation. Taku Komura Crowd simulation Taku Komura Animating Crowds We have been going through methods to simulate individual characters What if we want to simulate the movement of crowds? Pedestrians in the streets Flock of

More information

Computer Graphics I Lecture 11

Computer Graphics I Lecture 11 15-462 Computer Graphics I Lecture 11 Midterm Review Assignment 3 Movie Midterm Review Midterm Preview February 26, 2002 Frank Pfenning Carnegie Mellon University http://www.cs.cmu.edu/~fp/courses/graphics/

More information

Chapter 2. Related Work Approach to Avatars

Chapter 2. Related Work Approach to Avatars Chapter 2 Related Work The first section of this chapter discusses the classification of different approaches to avatar, and the distinct differences between them. The second section gives an overview

More information

Computer Kit for Development, Modeling, Simulation and Animation of Mechatronic Systems

Computer Kit for Development, Modeling, Simulation and Animation of Mechatronic Systems Computer Kit for Development, Modeling, Simulation and Animation of Mechatronic Systems Karol Dobrovodský, Pavel Andris, Peter Kurdel Institute of Informatics, Slovak Academy of Sciences Dúbravská cesta

More information

ENHANCING THE CONTROL AND PERFORMANCE OF PARTICLE SYSTEMS THROUGH THE USE OF LOCAL ENVIRONMENTS. Abstract

ENHANCING THE CONTROL AND PERFORMANCE OF PARTICLE SYSTEMS THROUGH THE USE OF LOCAL ENVIRONMENTS. Abstract ENHANCING THE CONTROL AND PERFORMANCE OF PARTICLE SYSTEMS THROUGH THE USE OF LOCAL ENVIRONMENTS Daniel O. Kutz Richard R. Eckert State University of New York at Binghamton Binghamton, NY 13902 Abstract

More information

Why study Computer Vision?

Why study Computer Vision? Why study Computer Vision? Images and movies are everywhere Fast-growing collection of useful applications building representations of the 3D world from pictures automated surveillance (who s doing what)

More information

Character Animation COS 426

Character Animation COS 426 Character Animation COS 426 Syllabus I. Image processing II. Modeling III. Rendering IV. Animation Image Processing (Rusty Coleman, CS426, Fall99) Rendering (Michael Bostock, CS426, Fall99) Modeling (Dennis

More information

ROBOT SENSORS. 1. Proprioceptors

ROBOT SENSORS. 1. Proprioceptors ROBOT SENSORS Since the action capability is physically interacting with the environment, two types of sensors have to be used in any robotic system: - proprioceptors for the measurement of the robot s

More information

Motion Editing with Data Glove

Motion Editing with Data Glove Motion Editing with Data Glove Wai-Chun Lam City University of Hong Kong 83 Tat Chee Ave Kowloon, Hong Kong email:jerrylam@cityu.edu.hk Feng Zou City University of Hong Kong 83 Tat Chee Ave Kowloon, Hong

More information

Shadows in the graphics pipeline

Shadows in the graphics pipeline Shadows in the graphics pipeline Steve Marschner Cornell University CS 569 Spring 2008, 19 February There are a number of visual cues that help let the viewer know about the 3D relationships between objects

More information

C O M P U T E R G R A P H I C S. Computer Animation. Guoying Zhao 1 / 66

C O M P U T E R G R A P H I C S. Computer Animation. Guoying Zhao 1 / 66 Computer Animation Guoying Zhao 1 / 66 Basic Elements of Computer Graphics Modeling construct the 3D model of the scene Rendering Render the 3D model, compute the color of each pixel. The color is related

More information

Volumetric Deformable Models for Simulation of Laparoscopic Surgery

Volumetric Deformable Models for Simulation of Laparoscopic Surgery Volumetric Deformable Models for Simulation of Laparoscopic Surgery S. Cotin y, H. Delingette y, J.M. Clément z V. Tassetti z, J. Marescaux z, N. Ayache y y INRIA, Epidaure Project 2004, route des Lucioles,

More information

Announcements. Ray tracer is due in five days you should have started by now or you re going to have a bad week. Missing file posted on the web page

Announcements. Ray tracer is due in five days you should have started by now or you re going to have a bad week. Missing file posted on the web page Announcements Ray tracer is due in five days you should have started by now or you re going to have a bad week Missing file posted on the web page I m sorry for canceling class on Tuesday... 1 Animation

More information

Nao Devils Dortmund. Team Description Paper for RoboCup Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann

Nao Devils Dortmund. Team Description Paper for RoboCup Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann Nao Devils Dortmund Team Description Paper for RoboCup 2017 Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann Robotics Research Institute Section Information Technology TU Dortmund University 44221 Dortmund,

More information

CS248. Game Mechanics

CS248. Game Mechanics CS248 Game Mechanics INTRODUCTION TOM WANG 2007 BS/MS CS KEY GAME MECHANICS * * * * * WORLD BUILDING CONTROLS CAMERA AI PERFORMANCE WORLD BUILDING WORLD BUILDING Set the atmosphere and tone of the game.

More information