Você está na página 1de 1

There are a wealth of unexplored interaction techniques in the eld of hand-gestural and pen interaction.

From the representation and design of custom novel gestures (suitable for complex language-sets), to the abstraction of gestures from observations. With the advancement of hand-gestural interactions in consumer devices, there have been a cry for research on different gestural and pen states and the implications of these states. Varying combinations of pen and touch interactions have also been explored in recent literature - many with the goal of streamling the ow of the workspace. There are two predominant design choices used to determine emerging gestures. These are; the closed-design, or individual-design of a gesture, and the abstraction of gestures through user observations in familiar contexts. In both (but moreso in the former), a steeper learning curve is usually required to progress transition novice to expert users. (Long,A.C.,et.al.,1999) suggests the design of a tool that aids the user in creating their own gestures. They note that their feedforward system (effectively letting the user know the accuracy of the machines comprehension of their gesture) was important to the user, and required better visualisation than a table format. Since this paper, commercial multitouchenabled devices have emerged that have contradicted some of the desires in users that this paper surveyed, namely a multitude of gestures for varying tasks, and the ability to create their custom gestures. However, a multitude of gesutres can certainly benet the communication of complex languages in narrower cases such as the creation of 3d models and sign-languages. Current User-created gestures are efcient for power-users, and has the potential to be collated - and aid in the design of new universal gestures. To learn such novel gestures, visualization in feedback and feedforward systems have been shown to be effective teachers (ShadowGuides, Octopocus). Such dynamic feedback can benet (e.g. voice) recognition systems. The observation of common interactions have been used to abstract gestures. (Hover Widgets) did this to test the accuracy of different pen-gestures, whilst (Ken Hinckley, Bill Buxton, et. al.) observed paper-notebook interactions to develop a scrapbook app that employed common functions such as page tearing / tracking, piling, etc. (Pen and Touch) advocates a pen writes, and touch manipulates interaction - which aligns with our traditional assignment of pen and touch interactions. Pen states have also been explored, resulting in interesting implications regarding hovering state (hover widgets), haptic feedback (haptic pen), and grip mode (H.Song,et.al.,2011). The combination of pen and touch interactions have also been explored, with unimodal /+ multimodal input, sequence of interactions, and bimanual input. Together, the design of gestures, pen states, and the combination of interactions, the workspace can be extended with both of localized and xed (menu bezel, Pen+Touch) user interfaces. This allows better workow as users acquire learnt and abstracted interaction paradigms more intuitively.

Você também pode gostar