Show simple item record

dc.contributor.authorJuarez‐Perez, A.en_US
dc.contributor.authorKallmann, M.en_US
dc.contributor.editorBenes, Bedrich and Hauser, Helwigen_US
dc.date.accessioned2020-05-22T12:24:42Z
dc.date.available2020-05-22T12:24:42Z
dc.date.issued2020
dc.identifier.issn1467-8659
dc.identifier.urihttps://doi.org/10.1111/cgf.13802
dc.identifier.urihttps://diglib.eg.org:443/handle/10.1111/cgf13802
dc.description.abstractWe present a new methodology for enabling virtual humans to autonomously detect and perform complex full‐body interactions with their environments. Given a parameterized walking controller and a set of motion‐captured example interactions, our method is able to detect when interactions can occur and to coordinate the detected upper‐body interaction with the walking controller in order to achieve full‐body mobile interactions in similar situations. Our approach is based on learning spatial coordination features from the example motions and on associating body‐environment proximity information to the body configurations of each performed action. Body configurations become the input to a regression system, which in turn is able to generate new interactions for different situations in similar environments. The regression model is capable of selecting, encoding and replicating key spatial strategies with respect to body coordination and management of environment constraints as well as determining the correct moment in time and space for starting an interaction. As a result, we obtain an interactive controller able to detect and synthesize coordinated full‐body motions for a variety of complex interactions requiring body mobility. Our results achieve complex interactions, such as opening doors and drawing in a wide whiteboard. The presented approach introduces the concept of learning interaction coordination models that can be applied on top of any given walking controller. The obtained method is simple and flexible, it handles the detection of possible interactions and is suitable for real‐time applications.en_US
dc.publisher© 2020 Eurographics ‐ The European Association for Computer Graphics and John Wiley & Sons Ltden_US
dc.subjecthuman simulation
dc.subjectanimation
dc.subjectmotion control
dc.subjectbehavioural animation
dc.subject• Computing methodologies → Procedural animation; Motion capture
dc.titleDetection and Synthesis of Full‐Body Environment Interactions for Virtual Humansen_US
dc.description.seriesinformationComputer Graphics Forum
dc.description.sectionheadersArticles
dc.description.volume39
dc.description.number1
dc.identifier.doi10.1111/cgf.13802
dc.identifier.pages274-286


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record