The Horizon Report 2012 had accurately forecast that gesture-based computing would be an important technology to watch out for. Although gamers are already familiar with Nintendo Wii and Microsoft’s Kinect system extend that to hand and arm motions, or body movement, these first-generation technologies were often clunky and had limited movement and mobility. Gesture-based technology (also known as motion control) is much more than just gaming as it promises to revolutionize the way we interact with computing technology.
As Leap Motion becomes reality using a small box the size of a matchbox to handle motion control technology, users can navigate on their screens by waving their hands in the air, and launch and play games on PCs without ever touching the keyboard or mouse. A sensor is placed on his desk in front of the screen and connects via USB. Once connected, gesture-based computing allows users to engage in virtual activities with motions and movements similar to what they would use in the real world, manipulating content intuitively.
What type of learning applications can gesture-based computing be useful for? In medicine, for example, gesture-based motion control enables virtual autopsy using a multi-touch table. Detailed CT scans can be created from a living (or deceased person) and transferred to the table where they are manipulated with gestures, allowing forensic scientists to examine a body, make virtual cross-sections, and view layers including skin, muscle, blood vessels, and bone. Can you imagine what libraries and museum collections can do by adopting gesture-based computing?
More Resources:
Neßelrath, R., Lu, C., Schulz, C. H., Frey, J., & Alexandersson, J. (2011). A Gesture Based System for Context–Sensitive Interaction with Smart Homes. In Ambient Assisted Living (pp. 209-219). Springer Berlin Heidelberg. [Link]
Johnson, L., Adams Becker, S., Cummins, M., Estrada, V., Freeman, A., & Ludgate, H. (2013). NMC Horizon Report: 2013 Higher Education Edition. [Link]
Maiti, A. (2013, February). Interactive remote laboratories with gesture based interface through microsoft kinect. In Remote Engineering and Virtual Instrumentation (REV), 2013 10th International Conference on (pp. 1-4). IEEE. [Link]
Mistry, P., & Maes, P. (2009, December). SixthSense: a wearable gestural interface. In ACM SIGGRAPH ASIA 2009 Sketches (p. 11). ACM. [Link]
No comments:
Post a Comment