BabyX Learns Her First Words

Saturday, August 23, 2014


 Artificial Intelligence
BabyX is an experimental computer generated psychobiological simulation of an infant which learns and interacts in real time. The software integrates realistic facial simulation with computational neuroscience models of neural systems involved in interactive behavior and learning.




BabyX is an interactive animated virtual infant prototype created by Dr. Mark Sager and his team at the Auckland Bioengineering Institute
Laboratory for Animate Technologies. The software is a computer generated psychobiological simulation under development and is an experimental vehicle incorporating computational models of basic neural systems involved in interactive behavior and learning.

"What we’re doing is making models of the face, but we’re driving them with models of the brain in order to build a system which creates its own expressions, its own emotions, and it can learn and react."


First introduced at a TEDx event last year, BabyX has grown into Version 3.0, and is now learning her first words (see video at top).

These models are embodied through advanced 3D computer graphics models of the face and upper body of an infant. The system can analyse video and audio inputs in real time to react to the caregiver’s or peer’s behavior using behavioral models. According to Sagar, BabyX is "A simulation of a brain driving the simulation of a face."

BabyX

"BabyX embodies many of the technologies we work on in the Laboratory and is under continuous development, in its neural models, sensing systems and also the realism of its real time computer graphics," claims the project's website.

“What we’re doing is making models of the face, but we’re driving them with models of the brain in order to build a system which creates its own expressions, its own emotions, and it can learn and react,” says Dr Sagar.

“We are taking models from theoretical neuroscience and using those to create animation; we’re integrating current theories of how emotional and behavioral systems work and we’re using those to create artificial nerve signals to drive the face of a digital baby. The neural systems driving the animation can be explored in real time, that is to say, you can watch the brain of BabyX.

BabyX Neural activity

Related articles
BabyX is modeled after Sagar’s own daughter, Francesca, and is a driven by computational models of brain activity. The system shows neural activity and emotions and reacts in real time to those around it. The reactions of the viewer and of the computerized baby are entirely dependent on what happens in real time between them.

Sagar previously worked as the Special Projects Supervisor at Weta Digital. He was involved with the creation of technology for the digital characters in blockbusters such Avatar, King Kong, and Spiderman 2.

His pioneering work in computer-generated faces was awarded with two consecutive Oscars at the 2010 and 2011 Sci-tech awards, a branch of the Academy Awards that recognizes movie science and technological achievements.


SOURCE  Auckland Bioengineering Institute
Laboratory for Animate Technologies

By 33rd SquareEmbed

0 comments:

Post a Comment