AR Puppet - Animated Agents in Mixed Reality Environments
Body and facial gestures as well as
speech are familiar and widely accepted means of human
communication. Animated characters, often with autonomous and
affective behavior, have proven to be useful in man-machine
communication since they are able to exploit and deliver
information through multimodal channels and thus engage the user
in a natural conversation. Autonomous agents have been actively
researched in recent years as an interface to computerized systems
bridging the communication gap between man and computer, and the
real and virtual world. Augmented Reality (AR) applications share
the same goal through enhancement of the real environment with
useful virtual information, where virtual objects appear to
coexist with the real world.
Quick links
AR Lego
 |
|
This project is an important pilot application for the use of AR
agents. It implements a machine maintenance scenario employing
two AR agents to educate an untrained user to assemble, test and
maintain machines composed of active (engines and sensors) and
passive (cogwheels, gears, frames) parts. The two agents are a
real, augmented LEGO Mindstorms robot and a virtual repairman.
The animated virtual repairman serves as an autonomous
telepointer and explains about the robot with expressive,
natural gestures in the vicinity of significant physical
locations during the construction steps. By monitoring the
attributes of the LEGO robot and the current status of assembly
application, character behavior is automatically generated for
the repairman, which is synchronized to the physical actions
triggered on the physical robot (turning engines on/off, setting
direction/power, querying various sensors, etc.).
|
Selected publication: |
I. Barakonyi, T. Psik, D. Schmalstieg, "Agents That Talk And Hit Back: Animated
Agents in Augmented Reality", Proc. of IEEE and ACM International
Symposium on Mixed and Augmented Reality 2004 (ISMAR'04), Arlington, VA, USA, pp. 141-150, November 2004. |
|
Downloads: |
PDF file
YouTube video |
Monkeybridge
 |
|
Lifelike animated characters have received significant attention in recent years.
Augmented Reality as a new a media demands broader acceptance, which can be achieved
by widening its application base to include entertainment and education. In this project
we show how animated agents make autonomous decisions based on their observation of an AR
environment in which they are embedded. Monkeybridge is a multiplayer game, where users
place real and virtual objects onto a physical surface, thus influencing the behavior of
the characters.The characters autonomously choose: the path the walk on; decide how to get
from one platform to the other, e.g. climb or jump when there is a slight difference in
height between platform edges; automatically choose the straightest path from several
available tiles; and fall into the water if there is no suitable piece of landing stage
to walk on. The game includes many spectacular virtual and physical visual elements such
as animated 3D ocean waves, a flock of virtual seagull boids, a physical volcano with real
smoke and a lighthouse with rotating lights. Sound effects enhance the game experience.
|
Selected publication: |
I. Barakonyi, M. Weilguny, T. Psik, D. Schmalstieg, "MonkeyBridge: Autonomous Agents in Augmented Reality Games",
Proc. of ACM SIGCHI International Conference on Advances in Computer Entertainment Technology (ACE'05), Valencia, Spain,
pp. 172-175, June 2005. |
|
Downloads: |
PDF file
YouTube video |
Animated agents in a UbiComp environment
 |
|
Most of today's Augmented Reality (AR) systems operate as passive information browsers
relying on a finite and deterministic world model and a predefined hardware and software
infrastructure. We propose an AR framework that dynamically and proactively exploits hitherto
unknown applications and hardware devices, and adapts the appearance of the user interface
to persistently stored and accumulated user preferences. Our framework explores proactive
computing, multi-user interface adaptation, and user interface migration. We employ mobile
and autonomous agents embodied by real and virtual objects as an interface and interaction
metaphor, where agent bodies are able to opportunistically migrate between multiple AR applications
and computing platforms to best match the needs of the current application context. We present two
pilot applications to illustrate design concepts.
|
Selected publication: |
I. Barakonyi, D. Schmalstieg, "Ubiquitous Animated Agents for Augmented Reality",
Proc. of IEEE and ACM International Symposium on Mixed and Augmented Reality 2006 (ISMAR'06),
Santa Barbara, CA, USA, pp. 145-154, October 2006. |
|
Downloads: |
PDF file
YouTube video |
AR calibration assistant for an ultra-wideband tracking system
 |
|
We have developed a hand-held augmented reality platform exploiting
a combination of multiple sensors built around an ultra-wideband tracking
system. We illustrate how an environment exploiting
this platform can be set up with a technician-support application, which
provides intuitive in-situ instructions on how a wide area tracking system
should be configured. The use of 3D registered graphics greatly assists in the
debugging of common awkward use cases involving reflections off metal surfaces.
|
Selected publication: |
J. Newman, G. Schall, I. Barakonyi, A. Schuerzinger, D. Schmalstieg,
"Sentient Environments for Augmented Reality", In Advances in Pervasive Computing,
Pervasive 2006, Dublin, Ireland, May 2006. |
|
Downloads: |
PDF file
YouTube video coming soon |
AR Piano Tutor
 |
|
In this project virtual objects are superimposed on top of a real MIDI keyboard.
The input and output channels of the keyboard are connected to a computer via a
MIDI interface. The application serves as a piano teacher that teaches about basic
chords, scales, etc. The keys pressed by the "student" user, the order and the
timing of the piano key presses are all captured by the computer with the help of
the MIDI interface, therefore the teacher can immediately "check" whether the student
has correctly followed the instructions and may give instant visual feedback over the
real keyboard (e.g. marking wrongly pressed or missed keys with different colors).
The application is based on a "fishtank" AR setup (PC+monitor+webcam), where the physical
MIDI keyboard is tracked with the help of an optical marker.
|
Selected publication: |
I. Barakonyi, D. Schmalstieg, "Augmented Reality Agents in the Development Pipeline
of Computer Entertainment", Proc. of 4th International Conference on Entertainment Computer
(ICEC'05), Sanda, Japan, pp. 345-356, September 2005. |
|
Downloads: |
PDF file
YouTube video |
Rotoscoping techniques exploiting AR
 |
|
Modellers and animators often rely on real-life references to build and animate
3D characters for games or film production. Observing the real world by photographing
or videotaping the subject, or asking someone to pose helps create more believable,
precise and expressive character animation. Since physical and virtual objects appear to coexist
in an Augmented Reality (AR) environment,
new possibilities in character animation open up. The animated virtual model and the real-world
reference can be merged into one, interactive modeling tool. The main goal of this project is to
develop such a tool using pose-tracked physical objects as input devices to animate
skeleton-based virtual 3D models. AR enables not only a close interaction with the virtual
model by using tangible objects but also the creation of complex motions like walking up stairs or
lifting a ball, since the animator can use the actual physical models of the stairs or the ball
together with the character to create the motion.
|
Selected publication: |
I. Barakonyi, D. Schmalstieg, "Augmented Reality in the Character Animation Pipeline",
ACM SIGGRAPH 2006 Sketches, Boston, MA, USA, July 2006. |
|
Downloads: |
PDF file
YouTube video |
AR videoconferencing
 |
|
The Augmented Reality Videoconferencing System is a novel remote collaboration tool
combining a desktop-based AR system and a videoconference module. The novelty of
this system is the combination of these tools with AR applications superimposed
on live video background displaying the conference parties' real environment,
merging the advantages of the natural face-to-face communication of videoconferencing
and AR’s interaction capabilities with distributed virtual objects using tangible
physical artifacts. The simplicity of the system makes it affordable for everyday use.
The system is based on concurrent video streaming, optical tracking and 3D application
sharing. The technology used provides yields superior quality compared to pure video
streaming with successive optical tracking from the compressed streams. The system's
collaborative features are demonstrated with a volume rendering application that allows
users to display and examine volumetric data simultaneously and to highlight or explore
slices of the volume by manipulating an optical marker as a cutting plane interaction device.
|
Selected publication: |
I. Barakonyi, T. Fahmy, D. Schmalstieg, "Remote Collaboration Using Augmented Reality Videoconferencing",
Proc. of Graphics Interface 2004 (GI'04), London, ON, Canada, pp. 89-96, May 2004. |
|
Downloads: |
PDF file
YouTube video |
Cascading Hand and Eye Movement for AR Videoconferencing
Research done at the National Institute of Informatics, Tokyo

 |
|
We have implemented an augmented reality videoconferencing
system that inserts virtual graphics overlays into the live video
stream of remote conference participants. The virtual objects are
manipulated using a novel interaction technique cascading
bimanual tangible interaction and eye tracking. User studies prove
that our user interface enriches remote collaboration by offering
hitherto unexplored ways for collaborative object manipulation
such as gaze controlled raypicking of remote physical and virtual
objects.
|
Selected publication: |
I. Barakonyi, H. Prendinger, D. Schmalstieg, M. Ishizuka, "Cascading hand and
eye movement for augmented reality videoconferencing", Proc. of 2nd IEEE Symposium on 3D User Interfaces 2007 (3DUI'07),
Charlotte, NC, USA, pp 71-78, March 2007. |
|
Downloads: |
PDF file
YouTube video |
Animated talking head for multimodal presentations
Research done at the Ishizuka Lab at the University of Tokyo

 |
|
This project presents a three-dimensional agent with synthesized face, speech and behavior.
The behavior synthesis is semiautonomous, which means the agent is capable of generating facial expressions,
head and eye motion automatically with a little user support of defining behavioral rules and annotating text
for the speech. Besides standard input and output devices we utilize speech synthesis and recognition for user
interaction. The agent is implemented as a scriptable, individual object with a well-defined interface,
which enables scripting from other applications. As the system uses external parameter files - locally stored
or distributed over the Internet -, authors can freely modify both the visual appearance and the behavior.
The described features allow for a versatile use of the agent, for instance performing multimodal presentations
without a human presenter, making conversations with the user or acting as an interactive teacher in distance learning.
|
Selected publication: |
N.P. Chandrasiri, I. Barakonyi, T. Naemura, M. Ishizuka, H. Harashima,
"Internet Communication Using Real-time Facial Expression Analysis and Synthesis",
IEEE Multimedia, Vol.11, No. 3, pp. 20-29, July-Sept 2004. |
|
Downloads: |
PDF file
YouTube video |
Publications
-
Alessandro Mulloni, Daniel Wagner, Istvan Barakonyi, Dieter Schmalstieg,
"Indoor Positioning and Navigation with Camera Phones",
In IEEE Pervasive Computing, Vol. 8, No. 2, pp. 22-31, April-June 2009.
Link to publisher
-
Dieter Schmalstieg, Gerhard Schall, Daniel Wagner, Istvan Barakonyi,
Gerhard Reitmayr, Joseph Newman, Florian Ledermann,
"Managing Complex Augmented Reality Models",
In IEEE Computer Graphics and Applications, Vol. 27, No. 4, pp. 48-57, July-August 2007.
Download (347Kb)
-
Istvan Barakonyi, Dieter Schmalstieg, Helmut Prendinger, Mitsuru Ishizuka,
"Cascading Hand and Eye Movement for Augmented Reality Videoconferencing",
In Proc. IEEE Symposium on 3D User Interfaces (3DUI'07), Charlotte NC, USA, March 2007.
Download (3.3Mb)
-
Istvan Barakonyi, Dieter Schmalstieg, "Ubiquitous Animated
Agents for Augmented Reality", In Proc. of IEEE and ACM
International Symposium on Mixed and Augmented Reality 2006
(ISMAR'06), Santa Barbara, CA, USA, October 2006.
Download (0.96Mb)
-
Istvan Barakonyi, Dieter Schmalstieg, "Augmented Reality in
the Character Animation Pipeline", SIGGRAPH sketch,
SIGGRAPH 2006, Boston, MA, USA, August 2006.
Download (2.8Mb)
-
Joseph Newman, Gerhard Schall, Istvan Barakonyi, Andreas
Schuerzinger, Dieter Schmalstieg, "Sentient Environments for
Augmented Reality", demo submission paper, In Advances in
Pervasive Computing, Pervasive 2006, Dublin, Ireland, May 2006.
Download (2.97Mb)
-
Florian Ledermann, Istvan Barakonyi, Dieter Schmalstieg, "Abstraction
and Implementation Strategies for Augmented Reality Authoring",
book chapter in Emerging Technologies of Augmented Reality:
Interfaces and Design, (M. Haller, B. Thomas, M. Billinghurst
eds.), Idea Group Publishing, to be published in 2006.
Download (231Kb)
-
Istvan Barakonyi, Dieter Schmalstieg, "Augmented Reality
Agents in the Development Pipeline of Computer Entertainment",
Proc. of the 4th International Conference on Entertainment
Computer (ICEC'05), Sanda, Japan, Sept. 19-21, 2005.
Download (2.6Mb)
-
Istvan Barakonyi, Markus Weilguny, Thomas Psik, Dieter
Schmalstieg, "MonkeyBridge: Autonomous Agents in Augmented
Reality Games", Proc. of the ACM SIGCHI International
Conference on Advances in Computer Entertainment Technology
(ACE'05), Valencia, Spain, June 15-17, 2005.
Download (1.3Mb)
-
Istvan Barakonyi, Dieter Schmalstieg, "Exploiting the
Physical World as User Interface in Augmented Reality
Applications", Proc. of the IEEE Virtual Reality 2005
Workshop on New Directions in 3D User Interfaces, Bonn, Germany,
March 12, 2005.
Download (1.6Mb)
-
Istvan Barakonyi, Thomas Psik, Dieter Schmalstieg, "Agents
That Talk And Hit Back: Animated Agents in Augmented Reality",
Proc. of the IEEE and ACM International Symposium on Mixed and
Augmented Reality 2004 (ISMAR'04), Arlington, VA, USA, Nov. 2-5,
2004, pp. 141-150.
Download (1.4Mb)
-
Istvan Barakonyi, Tamer Fahmy, Dieter Schmalstieg, "Remote
Collaboration Using Augmented Reality Videoconferencing",
Proc. of Graphics Interface 2004 (GI'04), London, ON, Canada,
May 17-19, 2004, pp. 89-96.
Download (701Kb)
-
Naiwala P. Chandrasiri, Istvan Barakonyi, Takeshi Naemura,
Mitsuru Ishizuka, Hiroshi Harashima, "Internet Communication
Using Real-time Facial Expression Analysis and Synthesis",
IEEE Multimedia, Vol.11, No. 3, July-Sept 2004, pp. 20-29.
Download (1.98Mb)
- Istvan Barakonyi, Dieter Schmalstieg, "AR Puppet: Animated
Agents in Augmented Reality ", Proc. of First Central
European International Multimedia and Virtual Reality Conference,
pp. 35-42, Veszprém, Hungary, May 6-8, 2004.
Download (295Kb)
-
Istvan Barakonyi, Tamer Fahmy, Dieter Schmalstieg, Karin Kosina,
"Collaborative Work with Volumetric Data Using Augmented
Reality Videoconferencing", Demo Paper, Proc. of the IEEE
and ACM International Symposium on Mixed and Augmented Reality
(ISMAR'03), Tokyo, Japan, October 2003.
Download (212Kb)
-
Istvan Barakonyi, Werner Frieb, Dieter Schmalstieg, "Augmented
Reality Videoconferencing for Collaborative Work", Proc. of
the 2nd Hungarian Conference on Computer Graphics and Geometry,
Budapest, Hungary, June 30 - July 1, 2003.
-
Naiwala P. Chandrasiri, Istvan Barakonyi, Takeshi Naemura,
Mitsuru Ishizuka and Hiroshi Harashima, "Communication over
the Internet using a 3D Agent with Real-time Facial Expression
Analysis, Synthesis and Text to Speech Capabilities", Proc.
of the IEEE International Conference on Communication Systems (ICCS
2002), Singapore, November 2002.
Download (399Kb)
-
Istvan Barakonyi, Naiwala P. Chandrasiri, Sylvain Descamps,
Mitsuru Ishizuka, "Communicating Multimodal Information on
the WWW Using a Lifelike, Animated 3D Agent", Proc. of the
PRICAI'02 workshop, Tokyo, Japan, August 2002 and the 1st
Hungarian Conference on Computer Graphics and Geometry,
Budapest, Hungary, May 2002.
Download (145Kb)
-
Sylvain Descamps, Istvan Barakonyi, Mitsuru Ishizuka, "Making
the Web Emotional: Authoring Multimodal Presentations Using a
Synthetic 3D Agent", Proc. of the OZCHI 2001 Conference,
Perth, Australia, November 2001, pp.25-30.
Download (246Kb)
-
Istvan Barakonyi, Mitsuru Ishizuka, "A 3D Agent with
Synthetic Face and Semiautonomous Behavior for Multimodal
Presentations", Proc. of the Multimedia Technology and
Applications Conference (MTAC2001, IEEE Computer Society),
Irvine, California, USA, November 2001, pp.21-25.
Download (247Kb)
|