User-defined Gestures for AR
We applied a guessibility study to obtain user-centered designed gestures for AR. Previous research had primarily focused on their proposed gestures, while little is known about user preferences and behavior interacting in AR.
We present the result of our study where 800 gestures were elicited for 40 selected tasks from 20 participants. Using the agreement found among gestures, a user-defined gesture set was created.
In term of research contributions, we extended Wobbrock’s surface taxonomy to cover dimensionalities in AR and with it, we derived the characteristics of collected gestures. We shared common motifs from this study for a better understanding of users’ thought and behavior.
We hope that our gesture set will assist designers in creating better and a more consistent user-centered gestures in AR.