Many status-quo interfaces for tablets with pen + touch input capabilities force users to reach for device-centric UI widgets at fixed locations, rather than sensing and adapting to the user-centric posture.
To address this problem, we propose sensing techniques that transition between various nuances of mobile and stationary use via postural awareness. These postural nuances include shifting hand grips, varying screen angle and orientation, planting the palm while writing or sketching, and detecting what direction the hands approach from.
To achieve this, our system combines three sensing modalities:
- raw capacitance touchscreen images,
- inertial motion, and
- electric field sensors around the screen bezel for grasp and hand proximity detection.
We show how these sensors enable posture-aware pen+touch techniques that adapt interaction and morph user interface elements to suit fine-grained contexts of body-, arm-, hand-, and grip-centric frames of reference.
(This video accompanies “Sensing Posture-Aware Pen+Touch Interaction on Tablets,” research conducted at Microsoft Research in Redmond, WA and published in the ACM CHI 2019 Conference on Human Factors in Computing Systems—where the work received an Honorable Mention award.)