3D Display of Virtual Humans. Kristina Khuu & Luana Sanchez MOVES Institute, Naval Postgraduate School, Monterey, CA 93943 Dr. Amela Sadagic, Research Associate Professor, Virtual Humans Lab. Abstract
Kristina Khuu & LuanaSanchezMOVES Institute, Naval Postgraduate School, Monterey, CA 93943
Dr. Amela Sadagic, Research Associate Professor, VirtualHumans Lab
3D virtual humans represent digital models of human beings. These models are used as an important resource for research studies, learning and training systems, video games, etc. In situations when the quality of interactions with virtual humans is as good as (or close to) the quality of interactions with real humans, 3D virtual humans can be used instead. 3D virtual humans provide a flexibility that may be needed in given system - they are available any time we 'ask' them and they represent an inexpensive alternative to hiring real humans. This project focused on revising a set of 3D virtual characters that are planned to be used for research studies at MOVES Institute. The effort included three distinct sets of activities: (1) identifying the flaws in visual appearance (3D geometry and textures), (2) testing the functionalities of the Face Controller application, and (3) building a comprehensive set of animation scenarios. The goal was to look for discoloration of skin, believability of emotions of facial expressions, the errors in the programing and to create a set of behaviors that would be used in user studies. Flaws found in the program were analyzed and logged in a report - each error found has been presented with the pictures and short description. This resource will later on be used by the programmers who will then know what needed to be fixed.
This project was focused on testing the models of 3D virtual humans with the Face Controller application. The logs of all errors identified will be used to fix and adjust both 3D models and the Face Controller application. The updated versions will be used for a series of user studies planned to be executed by researchers at the MOVES Institute. The studies are designed to help the researchers improve their understanding on how real humans react when presented with the situations in which they need to collaborate with virtual humans, and what characteristics of virtual humans need to be improved to make sure the interaction with virtual humans is qualified by the human participants as realistic and effective.
1. Revisions of Appearance
Figure 2. Example of discoloration on side of head.
2. Functionalities of the Face Controller Application
Figure 3. The figure at left illustrates a command and visual error. Polygons protrude out of his neck.
3. Animated Scenarios
A special thanks to Dr. AmelaSadagic and Professor Mathias Kolsch for being kind and nurturing mentors. A thanks to Alison Kerr of the Naval Postgraduate Cebrowski Institute, Joe Welch, Computer Science at Hartnell College, Andy Newton, Director of the Science Math Institute at Hartnell College. Lastly a thanks to everyone at the MOVES Institute for our success. The Internship was funded by the College to University Success Program.
I am looking forward to seeing you next week.
Figure 1. Example of a virtual reality head (left) and a true 3D virtual reality head with image projected on it (right).
Figure 4. The images above are screen shots of one of the animated scenarios and the tracks that control it.
For more information contact Dr. Amela Sadagic, email: firstname.lastname@example.org, Kristina Khuukgkhuu@gmail.com, Luana Sanchez, email@example.com
Figure 5. Example of a full body animated scenario and the tracks that control it.