This new way of interaction is especially suitable for wearables because it uses only our own body for both input and output and is eyes-free. It is based on the user’s proprioceptive sense, i.e., rather than seeing, hearing, or feeling an outside stimulus, users feel the pose of their own body. To pose the users body (output) our devices use electrical muscle stimulation, as it allows us to actuate with a minimal hardware footprint (mobile & wearable). For controlling the interaction users input by posing their body, which we read with different tracking methods (optical, accelerometer, etc). Here’s some examples of the work I have done on proprioceptive interaction: