Jump to content

Hand tracking: Difference between revisions

No edit summary
Line 23: Line 23:


=== Early Developments (1970s–1990s) ===
=== Early Developments (1970s–1990s) ===
The foundational milestone occurred in 1977 with the invention of the '''[[Sayre Glove]]''', a wired data glove developed by electronic visualization pioneer Daniel Sandin and computer graphics researcher Thomas DeFanti at the University of Illinois at Chicago's Electronic Visualization Laboratory (EVL). Inspired by an idea from colleague Rich Sayre, the glove used optical flex sensors—light emitters paired with photocells embedded in the fingers—to measure joint angles and finger bends. Light intensity variations were converted into electrical signals, enabling basic gesture recognition and hand posture tracking for early VR simulations.<ref name="SayreGlove" /><ref name="SenseGlove" /> This device, considered the first data glove, established the principle of measuring finger flexion for computer input.
The foundational milestone occurred in 1977 with the invention of the '''[[Sayre Glove]]''', a wired data glove developed by electronic visualization pioneer Daniel Sandin and computer graphics researcher Thomas DeFanti at the University of Illinois at Chicago's Electronic Visualization Laboratory (EVL). Inspired by an idea from colleague Rich Sayre, the glove used optical flex sensors, light emitters paired with photocells embedded in the fingers, to measure joint angles and finger bends. Light intensity variations were converted into electrical signals, enabling basic gesture recognition and hand posture tracking for early VR simulations.<ref name="SayreGlove" /><ref name="SenseGlove" /> This device, considered the first data glove, established the principle of measuring finger flexion for computer input.


In 1983, Gary Grimes of Bell Labs developed the '''[[Digital Data Entry Glove]]''', a more sophisticated system patented as an alternative to keyboard input. This device integrated flex sensors, touch sensors, and tilt sensors to recognize unique hand positions corresponding to alphanumeric characters, specifically gestures from the American Sign Language manual alphabet.<ref name="BellGlove" />
In 1983, Gary Grimes of Bell Labs developed the '''[[Digital Data Entry Glove]]''', a more sophisticated system patented as an alternative to keyboard input. This device integrated flex sensors, touch sensors, and tilt sensors to recognize unique hand positions corresponding to alphanumeric characters, specifically gestures from the American Sign Language manual alphabet.<ref name="BellGlove" />
Line 52: Line 52:
The 2000s saw the convergence of hardware and software for multi-modal tracking. External devices like data gloves with fiber-optic sensors (e.g., Fifth Dimension Technologies' 5DT Glove) combined bend sensors with IMUs to capture 3D hand poses. Software frameworks began processing fused data for virtual hand avatars. However, these remained bulky and controller-dependent, with limited adoption outside research labs.<ref name="VirtualSpeech" />
The 2000s saw the convergence of hardware and software for multi-modal tracking. External devices like data gloves with fiber-optic sensors (e.g., Fifth Dimension Technologies' 5DT Glove) combined bend sensors with IMUs to capture 3D hand poses. Software frameworks began processing fused data for virtual hand avatars. However, these remained bulky and controller-dependent, with limited adoption outside research labs.<ref name="VirtualSpeech" />


In the late 1990s and early 2000s, camera-based gesture recognition began to be explored outside of VR—for instance, computer vision researchers worked on interpreting hand signs for sign language or basic gesture control of computers. However, real-time markerless hand tracking in 3D was extremely challenging with the processing power then available.
In the late 1990s and early 2000s, camera-based gesture recognition began to be explored outside of VR, for instance, computer vision researchers worked on interpreting hand signs for sign language or basic gesture control of computers. However, real-time markerless hand tracking in 3D was extremely challenging with the processing power then available.


=== 2010s: Optical Tracking and Controller-Free Era ===
=== 2010s: Optical Tracking and Controller-Free Era ===
Line 76: Line 76:
In the 2020s, hand tracking became an expected feature in many XR devices. An analysis by SpectreXR noted that the percentage of new VR devices supporting hand tracking jumped from around 21% in 2021 to 46% in 2022, as more manufacturers integrated the technology.<ref name="SpectreXR2023" /> At the same time, the cost barrier dropped dramatically, with the average price of hand-tracking-capable VR headsets falling by approximately 93% from 2021 to 2022, making the technology far more accessible.<ref name="SpectreXR2023" />
In the 2020s, hand tracking became an expected feature in many XR devices. An analysis by SpectreXR noted that the percentage of new VR devices supporting hand tracking jumped from around 21% in 2021 to 46% in 2022, as more manufacturers integrated the technology.<ref name="SpectreXR2023" /> At the same time, the cost barrier dropped dramatically, with the average price of hand-tracking-capable VR headsets falling by approximately 93% from 2021 to 2022, making the technology far more accessible.<ref name="SpectreXR2023" />


Another milestone was Apple's introduction of the '''[[Apple Vision Pro]]''' (released 2024), which relies on hand tracking along with [[eye tracking]] as the primary input method for a spatial computer, completely doing away with handheld controllers. Apple's implementation allows users to make micro-gestures like pinching fingers at waist level, tracked by downward-facing cameras, which—combined with eye gaze—lets users control the interface in a very effortless manner.<ref name="AppleGestures" /><ref name="UploadVR2023" /> This high-profile adoption has been seen as a strong endorsement of hand tracking for mainstream XR interaction.
Another milestone was Apple's introduction of the '''[[Apple Vision Pro]]''' (released 2024), which relies on hand tracking along with [[eye tracking]] as the primary input method for a spatial computer, completely doing away with handheld controllers. Apple's implementation allows users to make micro-gestures like pinching fingers at waist level, tracked by downward-facing cameras, which, combined with eye gaze, lets users control the interface in a very effortless manner.<ref name="AppleGestures" /><ref name="UploadVR2023" /> This high-profile adoption has been seen as a strong endorsement of hand tracking for mainstream XR interaction.


By 2025, hand tracking is standard in many XR devices, with latencies under 70 ms and applications spanning gaming to medical simulations.
By 2025, hand tracking is standard in many XR devices, with latencies under 70 ms and applications spanning gaming to medical simulations.
Line 173: Line 173:
=== Multimodal Interaction ===
=== Multimodal Interaction ===
Combining hand tracking with other inputs enhances interaction:
Combining hand tracking with other inputs enhances interaction:
* '''[[Gaze-and-pinch]]''' ([[Apple Vision Pro]]): [[Eye tracking]] rapidly targets UI elements, while a subtle pinch gesture confirms selection. This is the primary paradigm on [[Apple Vision Pro]], allowing control without holding up hands constantly—a brief pinch at waist level suffices.<ref name="AppleGestures" /><ref name="UploadVR2023" />
* '''[[Gaze-and-pinch]]''' ([[Apple Vision Pro]]): [[Eye tracking]] rapidly targets UI elements, while a subtle pinch gesture confirms selection. This is the primary paradigm on [[Apple Vision Pro]], allowing control without holding up hands constantly, a brief pinch at waist level suffices.<ref name="AppleGestures" /><ref name="UploadVR2023" />
* '''[[Voice]] and [[gesture]]''': Verbal commands with hand confirmation
* '''[[Voice]] and [[gesture]]''': Verbal commands with hand confirmation
* '''Hybrid controller/hands''': Seamless switching between modalities
* '''Hybrid controller/hands''': Seamless switching between modalities


=== Gesture Commands ===
=== Gesture Commands ===
Beyond direct object manipulation, hand tracking can facilitate recognition of symbolic gestures that act as commands. This is analogous to how touchscreens support multi-touch gestures (pinch to zoom, swipe to scroll). In XR, certain hand poses or movements can trigger actions—for example, making a pinching motion can act as a click or selection, a thumbs-up might trigger an event, or specific sign language gestures could be interpreted as system commands.
Beyond direct object manipulation, hand tracking can facilitate recognition of symbolic gestures that act as commands. This is analogous to how touchscreens support multi-touch gestures (pinch to zoom, swipe to scroll). In XR, certain hand poses or movements can trigger actions, for example, making a pinching motion can act as a click or selection, a thumbs-up might trigger an event, or specific sign language gestures could be interpreted as system commands.


=== User Interface Navigation ===
=== User Interface Navigation ===
Line 188: Line 188:
* '''System UI & Productivity''': Controller-free navigation, window management, and typing/pointing surrogates in spatial desktops. Natural file manipulation, multitasking across virtual screens, and interface control without handheld devices.<ref name="AppleGestures" />
* '''System UI & Productivity''': Controller-free navigation, window management, and typing/pointing surrogates in spatial desktops. Natural file manipulation, multitasking across virtual screens, and interface control without handheld devices.<ref name="AppleGestures" />


* '''Gaming & Entertainment''': Titles such as ''Hand Physics Lab'' showcase free-hand puzzles and physics interactions using optical hand tracking on Quest.<ref name="HPL_RoadToVR" /> Games and creative applications use hand interactions—e.g., a puzzle game might let the player literally reach out and grab puzzle pieces in VR, or users can play virtual piano or create pottery simulations.
* '''Gaming & Entertainment''': Titles such as ''Hand Physics Lab'' showcase free-hand puzzles and physics interactions using optical hand tracking on Quest.<ref name="HPL_RoadToVR" /> Games and creative applications use hand interactions, e.g., a puzzle game might let the player literally reach out and grab puzzle pieces in VR, or users can play virtual piano or create pottery simulations.


* '''Training & Simulation''': Natural hand use improves ecological validity for assembly, maintenance, and surgical rehearsal in enterprise, medical, and industrial contexts.<ref name="Frontiers2021" /> Workers can practice complex procedures in safe virtual environments, developing muscle memory that transfers to real-world tasks.
* '''Training & Simulation''': Natural hand use improves ecological validity for assembly, maintenance, and surgical rehearsal in enterprise, medical, and industrial contexts.<ref name="Frontiers2021" /> Workers can practice complex procedures in safe virtual environments, developing muscle memory that transfers to real-world tasks.
Line 194: Line 194:
* '''Social and Collaborative VR''': In multi-user virtual environments, hand tracking enhances communication and embodiment. Subtle hand motions and finger movements can be transmitted to one's avatar, allowing for richer non-verbal communication such as waving, pointing things out to others, or performing shared gestures. This mirrors real life and can make remote collaboration or socializing feel more natural.<ref name="VarjoSupport" />
* '''Social and Collaborative VR''': In multi-user virtual environments, hand tracking enhances communication and embodiment. Subtle hand motions and finger movements can be transmitted to one's avatar, allowing for richer non-verbal communication such as waving, pointing things out to others, or performing shared gestures. This mirrors real life and can make remote collaboration or socializing feel more natural.<ref name="VarjoSupport" />


* '''Accessibility & Rehabilitation''': Because hand tracking removes the need to hold controllers, it can make VR and AR more accessible to people who may not be able to use standard game controllers. Users with certain physical disabilities or limited dexterity might find hand gestures easier. In addition, the technology has been explored for rehabilitation exercises—for example, stroke patients could do guided therapy in VR using their hands to perform tasks and regain motor function, with the system tracking their movements and providing feedback. Reduces dependence on handheld controllers in shared or constrained environments.<ref name="Frontiers2021" />
* '''Accessibility & Rehabilitation''': Because hand tracking removes the need to hold controllers, it can make VR and AR more accessible to people who may not be able to use standard game controllers. Users with certain physical disabilities or limited dexterity might find hand gestures easier. In addition, the technology has been explored for rehabilitation exercises, for example, stroke patients could do guided therapy in VR using their hands to perform tasks and regain motor function, with the system tracking their movements and providing feedback. Reduces dependence on handheld controllers in shared or constrained environments.<ref name="Frontiers2021" />


* '''Healthcare & Medical''': AR HUD (heads-up display) interactions in medical contexts allow surgeons to manipulate virtual panels without touching anything physically, maintaining sterile fields. Medical training simulations benefit from realistic hand interactions.
* '''Healthcare & Medical''': AR HUD (heads-up display) interactions in medical contexts allow surgeons to manipulate virtual panels without touching anything physically, maintaining sterile fields. Medical training simulations benefit from realistic hand interactions.
Line 218: Line 218:
* '''Enhanced Immersion''': Removing intermediary devices (like controllers or wands) can increase presence. When users see their virtual hands mimicking their every finger wiggle, it reinforces the illusion that they are "inside" the virtual environment. The continuity between real and virtual actions (especially in MR, where users literally see their physical hands interacting with digital objects) can be compelling.
* '''Enhanced Immersion''': Removing intermediary devices (like controllers or wands) can increase presence. When users see their virtual hands mimicking their every finger wiggle, it reinforces the illusion that they are "inside" the virtual environment. The continuity between real and virtual actions (especially in MR, where users literally see their physical hands interacting with digital objects) can be compelling.


* '''Expressiveness''': Hands allow a wide range of gesture expressions. In contrast to a limited set of controller buttons, hand tracking can capture nuanced movements. This enables richer interactions (such as sculpting a 3D model with complex hand movements) and communication (subtle social gestures, sign language, etc.). Important for social presence—waving, pointing, subtle finger cues enhance non-verbal communication.
* '''Expressiveness''': Hands allow a wide range of gesture expressions. In contrast to a limited set of controller buttons, hand tracking can capture nuanced movements. This enables richer interactions (such as sculpting a 3D model with complex hand movements) and communication (subtle social gestures, sign language, etc.). Important for social presence, waving, pointing, subtle finger cues enhance non-verbal communication.


* '''Hygiene & Convenience''': Especially in public or shared XR setups, hand tracking can be advantageous since users do not need to touch common surfaces or devices. Touchless interfaces have gained appeal for reducing contact points. Moreover, not having to pick up or hold hardware means quicker setup and freedom to use one's hands spontaneously (e.g., switching between real objects and virtual interface by just moving hands). No shared controllers required; quicker task switching between physical tools and virtual UI.
* '''Hygiene & Convenience''': Especially in public or shared XR setups, hand tracking can be advantageous since users do not need to touch common surfaces or devices. Touchless interfaces have gained appeal for reducing contact points. Moreover, not having to pick up or hold hardware means quicker setup and freedom to use one's hands spontaneously (e.g., switching between real objects and virtual interface by just moving hands). No shared controllers required; quicker task switching between physical tools and virtual UI.
Line 260: Line 260:


=== Neural Networks for Better Prediction ===
=== Neural Networks for Better Prediction ===
There is active research into using neural networks for better prediction of occluded or fast movements, and into augmenting hand tracking with other sensors (for example, using electromyography—reading muscle signals in the forearm—to detect finger movements even before they are visible). All these efforts point toward making hand-based interaction more seamless, reliable, and richly interactive in the coming years.
There is active research into using neural networks for better prediction of occluded or fast movements, and into augmenting hand tracking with other sensors (for example, using electromyography, reading muscle signals in the forearm, to detect finger movements even before they are visible). All these efforts point toward making hand-based interaction more seamless, reliable, and richly interactive in the coming years.


=== Market Projections ===
=== Market Projections ===
Line 306: Line 306:
<ref name="Orion">TechNewsWorld – "Leap Motion Unleashes Orion" (2016-02-18). URL: https://www.technewsworld.com/story/leap-motion-unleashes-orion-83129.html</ref>
<ref name="Orion">TechNewsWorld – "Leap Motion Unleashes Orion" (2016-02-18). URL: https://www.technewsworld.com/story/leap-motion-unleashes-orion-83129.html</ref>
<ref name="UltrahapticsAcq">The Verge – "Hand-tracking startup Leap Motion reportedly acquired by UltraHaptics" (2019-05-30). URL: https://www.theverge.com/2019/5/30/18645604/leap-motion-vr-hand-tracking-ultrahaptics-acquisition-rumor</ref>
<ref name="UltrahapticsAcq">The Verge – "Hand-tracking startup Leap Motion reportedly acquired by UltraHaptics" (2019-05-30). URL: https://www.theverge.com/2019/5/30/18645604/leap-motion-vr-hand-tracking-ultrahaptics-acquisition-rumor</ref>
<ref name="Meta2019">Meta – "Introducing Hand Tracking on Oculus Quest—Bringing Your Real Hands into VR" (2019). URL: https://www.meta.com/blog/introducing-hand-tracking-on-oculus-quest-bringing-your-real-hands-into-vr/</ref>
<ref name="Meta2019">Meta – "Introducing Hand Tracking on Oculus Quest-Bringing Your Real Hands into VR" (2019). URL: https://www.meta.com/blog/introducing-hand-tracking-on-oculus-quest-bringing-your-real-hands-into-vr/</ref>
<ref name="SpectreXR2022">SpectreXR Blog – "Brief History of Hand Tracking in Virtual Reality" (Sept 7, 2022). URL: https://spectrexr.io/blog/news/brief-history-of-hand-tracking-in-virtual-reality</ref>
<ref name="SpectreXR2022">SpectreXR Blog – "Brief History of Hand Tracking in Virtual Reality" (Sept 7, 2022). URL: https://spectrexr.io/blog/news/brief-history-of-hand-tracking-in-virtual-reality</ref>
<ref name="Develop3D2019">Develop3D – "First Look at HoloLens 2" (Dec 20, 2019). URL: https://develop3d.com/reviews/first-look-hololens-2-microsoft-mixed-reality-visualisation-hmd/</ref>
<ref name="Develop3D2019">Develop3D – "First Look at HoloLens 2" (Dec 20, 2019). URL: https://develop3d.com/reviews/first-look-hololens-2-microsoft-mixed-reality-visualisation-hmd/</ref>