RealityTalk: Real-time Speech-driven Augmented Presentation for AR Live Storytelling

RealityTalk: Real-time Speech-driven Augmented Presentation for AR Live Storytelling

Jian Liao, Adnan Karim, Shivesh Jadon, Rubaiat Habib Kazi, Ryo Suzuki  

The ACM Symposium on User Interface Software and Technology (UIST 2022)

Links:     PDF   Video   ACM DL   arXiv


Abstract

We present RealityTalk, a system that augments real-time live presentations with speech-driven interactive virtual elements. Augmented presentations leverage embedded visuals and animation for engaging and expressive storytelling. However, existing tools for live presentations often lack interactivity and improvisation, while creating such effects in video editing tools require significant time and expertise. RealityTalk enables users to create live augmented presentations with real-time speech-driven interactions. The user can interactively prompt, move, and manipulate graphical elements through real-time speech and supporting modalities. Based on our analysis of 177 existing video-edited augmented presentations (https://ilab.ucalgary.ca/realitytalk/), we propose a novel set of interaction techniques and then incorporated them into RealityTalk. We evaluate our tool from a presenter's perspective to demonstrate the effectiveness of our system.

Publication

Jian Liao, Adnan Karim, Shivesh Jadon, Rubaiat Habib Kazi, and Ryo Suzuki. 2022. RealityTalk: Real-time Speech-driven Augmented Presentation for AR Live Storytelling. In Proceedings of the 35th Annual ACM Symposium on User Interface Software and Technology (UIST '22). ACM, New York, NY, USA, .
DOI: https://doi.org/10.1145/3526113.3545702


   Download PDF

Slide

coming soon