My research lies at the intersection of human-computer interaction (HCI), computer-supported cooperative work (CSCW), computer-mediated communication (CMC), and educational technology.
My primary research focus is to build rich collaboration systems that offer expressive multi-modal interactions, i.e., interactions through multiple communication channels (e.g., speech, gesture, and grasp). My design approach translates natural human interactions into novel combinations of input modalities that serve as building blocks for fluid, rich, and lightweight interfaces. My evaluation approach deploys and evaluates high-fidelity systems in real world contexts (e.g., classrooms), from which we can obtain ecologically valid user data.
Recently, I am focusing on deploying RichReview, a multi-modal collaborative annotation system, to university classrooms and massive open online courses (MOOCs). My PhD study is supported in part by Kwanjeong Scholarship.
CSCW 2016, UIST 2014
A multi-modal annotation system supports rich collaboration through voice + gesture expressions.
A transcription-based voice editing system reduces workloads in speech commenting.
Grasp + Micro-mobility
Employing capacitive grip sensing and inertial motion can capture contexts of collaborative interactions.
CHI WIP 2013
A multi-touch gesture supports nonlinear page navigation through a set of lightweight bookmarking interactions.
Using the smartphone as a 3D controller supports direct manipulation of an on-screen virtual object.
TypeTalker: A Speech Synthesis-Based Multi-Modal Commenting System
CSCW 2017 Full Paper Accepted
SimpleSpeech: Simplified Audio Production in Asynchronous Voice-Based Discussions
RichReview++: Deployment of a Collaborative Multi-Modal Annotation System for Instructor Feedback and Peer Discussion
Sensing Tablet Grasp + Micro-mobility for Active Reading
RichReview: blending ink, speech, and gesture to support collaborative document review
TextTearing: opening white space for digital ink annotation
Mobiature: 3d model manipulation technique for large displays using mobile devices
Enriching Online Classroom Communication with Collaborative Multimodal Annotations
UIST 2015 Doctoral Symposium paper
Multimodal Peer Discussion with RichReview on edX
UIST 2015 Demonstrations paper
Supporting Face-to-Face Like Communication Modalities for Asynchronous Assignment Feedback in Math Education
Learning at Scale 2015 Works-in-Progress paper
Let me show you what I read: exploring referencing strategies for e-books
SIGCHI 2013 Works-in-Progress paper
Touch-Bookmark: a lightweight navigation and bookmarking technique for e-books
edX, Research Intern (2015)
Researches on multi-modal peer-discussion at scale, mentored by Piotr Mitros
Microsoft Research, Research Intern (2014)
Researches on tablet grasp interactions, mentored by Ken Hinckley
Microsoft Research, Research Intern (2013)
Korea Institute of Science and Technology, Research Scientist (2009 - 2012)
Researches on multi-touch and augmented reality technologies and their usability issues
Xenix Studio, Developer (2010)
Software development and UI consulting for a cross platform mobile augmented reality software engine
Flur, Developer (2008 - 2009)
Graphics software development for media art exhibitions