Jump to content

Passthrough: Difference between revisions

From VR & AR Wiki
No edit summary
No edit summary
Line 1: Line 1:
{{stub}}
{{featured}} <!-- Assuming this combined, detailed version aims for featured status -->
[[Passthrough]], also known as "video passthrough", is a capability of mixed reality/virtual reality headsets where they can display the outside world through cameras to the user in VR. This makes it a sort of pseudo-augmented reality.
'''Passthrough''', often referred to as '''video passthrough''', is a feature found in [[Virtual Reality]] (VR) and [[Mixed Reality]] (MR) [[head-mounted display|headsets]] that utilizes external [[camera|cameras]] to capture a live video feed of the physical environment around the user and display it on the internal screens within the headset.<ref name="XRToday_def">XR Today – What is VR Passthrough and How is it Shaping the Future of XR? (Immersive Learning News, Dec 2024)</ref><ref name="VIVE_Blog_Sauce">VIVE Blog: What is VR Passthrough? Mixed Reality's Secret Sauce</ref> This capability effectively allows users to see the "real world" without removing the headset, bridging the gap between fully immersive virtual experiences and the user's actual surroundings.


One of the difficulties of getting this right is the positional difference between cameras on the outside of a headset and the eyes of the user. So the image has to go through a distortion process. On the [[Quest Pro]] and [[Quest 3]], this uses a neural network. (and potentially other devices).
While primarily a feature of VR headsets aiming to add environmental awareness or MR capabilities, it functions as a form of [[Augmented Reality]] (AR), often termed "Video See-Through AR" (VST AR) or sometimes "pseudo-AR," as opposed to "[[Optical See-Through]] AR" (OST AR) systems which use transparent displays.<ref name="SkarbezVSTvsOST">[https://www.researchgate.net/publication/315722770_Revisiting_Milgram_and_Kishino%27s_Reality-Virtuality_Continuum Revisiting Milgram and Kishino's Reality-Virtuality Continuum] - Discusses the spectrum including Video See-Through.</ref> Passthrough is a key enabler of [[mixed reality]] and [[spatial computing]] experiences on modern headsets.
 
== Core Technology and How It Works ==
The fundamental principle of passthrough involves a real-time processing pipeline:
 
# '''Capture:''' One or more outward-facing digital cameras mounted on the headset capture video of the external world. Early or basic systems might use a single camera (providing a monoscopic view), while more advanced systems use two or more cameras to capture [[stereoscopic]] video, enabling [[depth perception]].<ref name="StereoPassthrough">[https://ieeexplore.ieee.org/document/9191148 Example paper discussing stereoscopic passthrough challenges]</ref> Modern systems often use a combination of [[RGB]] color cameras and monochrome (grayscale) sensors for different purposes (e.g., capturing color data vs. motion/detail).<ref name="MixedNews_Cambria">MIXED News – Project Cambria: Meta explains new passthrough technology (Tomislav Bezmalinović, May 16, 2022)</ref>
# '''Processing:''' The captured video footage is sent to the headset's [[processor]] (either an onboard [[System on a Chip|SoC]] or a connected PC's [[GPU]]). This stage is computationally intensive and critical for a usable and comfortable experience. It typically involves several steps:
#* '''Rectification/Undistortion:''' Correcting [[lens distortion]] inherent in the wide-angle cameras typically used to maximize [[field of view|FOV]].
#* '''Reprojection/Warping:''' Adjusting the captured image perspective to align with the user's eye position inside the headset, rather than the camera's physical position on the outside. This difference in viewpoint causes [[parallax]], and correcting it ("perspective correction") is crucial for accurate spatial representation, correct scale perception, and minimizing [[motion sickness]].<ref name="PassthroughChallengesUploadVR">[https://uploadvr.com/passthrough-ar-technical-challenges/ Passthrough AR: The Technical Challenges of Blending Realities] - UploadVR article discussing latency, distortion, etc.</ref><ref name="KGuttag_Align">KGOnTech (Karl Guttag) – Perspective Correct Passthrough (Sept 26, 2023)</ref> Algorithms based on [[Computer Vision]] and potentially [[Inertial Measurement Unit|IMU]] sensor data are used. Some modern headsets, like the [[Meta Quest Pro]] and [[Meta Quest 3]], employ [[Machine Learning]] or [[Neural Network|neural networks]] to improve the realism and accuracy of this reconstruction.<ref name="QuestProPassthrough">[https://www.meta.com/blog/quest/meta-reality-passthrough-quest-pro/ Meta Blog: Inside Meta Reality and Passthrough on Quest Pro]</ref>
#* '''[[Sensor Fusion]]:''' Combining data from multiple cameras (e.g., fusing monochrome detail with RGB color<ref name="MixedNews_Cambria"/>) and integrating tracking data (e.g., from [[inside-out tracking]] sensors or [[depth sensor]]s) to ensure the passthrough view remains stable, depth-correct, and aligned with the user's head movements.
#* '''Color Correction & Enhancement:''' Adjusting colors, brightness, and contrast to appear more natural, especially under varying lighting conditions. This can also involve [[Artificial Intelligence|AI]]-based denoising or upscaling.<ref name="UploadVR_Q3Review">UploadVR – Quest 3 Review: Excellent VR With Limited MR (David Heaney, Oct 9, 2023)</ref>
# '''Display:''' The processed video feed is rendered onto the headset's internal [[display|displays]], replacing or being overlaid upon the virtual content. The primary goal is to achieve this entire pipeline with minimal [[latency (engineering)|latency]] (ideally under 20 milliseconds<ref name="LatencyThreshold">[https://research.nvidia.com/publication/2016-07_Latency-Requirements-Plausible-Interaction-Augmented-and-Virtual-Reality Latency Requirements for Plausible Interaction in Augmented and Virtual Reality] - Research discussing latency impact.</ref>) to avoid discomfort and maintain realism.
 
== History and Evolution ==
While the concept of video passthrough existed in research labs for decades,<ref name="MilgramKishino1994">Milgram, P., & Kishino, F. (1994). A taxonomy of mixed reality visual displays. IEICE Transactions on Information Systems, E77-D(12), 1321-1329.</ref> its implementation in consumer VR headsets evolved significantly:
 
*'''Early Stages (Mid-2010s):''' Passthrough began appearing primarily as a safety feature. In 2016, the [[HTC Vive]] prototype (Vive Pre) introduced a front-facing camera providing a basic, monochrome, 2D view for obstacle avoidance. Valve's software projected this onto a virtual sphere to approximate perspective.<ref name="RoadToVR_Vive">Road to VR – 8 Minutes of the HTC Vive’s Front-facing Camera in Action (Paul James, Mar 10, 2016)</ref> It was low-resolution and intended for brief checks.
*'''Integrated Monochrome (Late 2010s):''' Headsets using inside-out tracking leveraged their tracking cameras for improved passthrough. The [[Oculus Rift S]] (2019) offered "Passthrough+" using its multiple monochrome cameras for a stereoscopic view.<ref name="RiftS_Docs">Oculus Rift S Product Documentation. (2019).</ref> The original [[Meta Quest|Oculus Quest]] (2019) and [[Meta Quest 2]] (2020) provided similar basic monochrome passthrough, mainly for setting up the [[Guardian system]] and quick environment checks.<ref name="Quest2Passthrough">[https://www.meta.com/blog/quest/oculus-quest-2-passthrough-public-api-passthrough/ Quest Blog on Quest 2 Passthrough improvements]</ref>
*'''Early Mixed Reality Steps (Early 2020s):''' In 2021, Meta released an experimental Passthrough API for Quest 2 developers, allowing apps to overlay virtual elements onto the monochrome feed, marking a step towards consumer MR.<ref name="PCMag_passthrough">PCMag – Oculus Experiments With Mixed Reality via New Passthrough API (Nathaniel Mott, July 25, 2021)</ref> Simultaneously, enterprise headsets like the [[Varjo]] XR-1 (2019) and XR-3 (2021) pushed high-fidelity color passthrough with dual high-resolution cameras, setting a benchmark for quality.<ref name="Skarredghost_Varjo">The Ghost Howls – Varjo XR-3 hands-on review (Tony Vitillo, June 8, 2022)</ref>
*'''Mainstream Color Passthrough (2022-Present):'''
    * The [[Meta Quest Pro]] (2022) was the first major consumer headset featuring high-quality, stereoscopic color passthrough, using a novel camera array (monochrome for depth/detail, RGB for color) and ML reconstruction.<ref name="MixedNews_Cambria"/>
    * Competitors like the [[Pico 4]] (late 2022) and [[HTC Vive XR Elite]] (2023) also introduced color passthrough, although early implementations like the Pico 4's were initially monoscopic and lacked depth correction.<ref name="Reddit_PicoMono">Reddit – Meta Quest Pro vs PICO 4 Passthrough Comparison (discussion summary, 2022)</ref><ref name="ViveXRElite">[https://www.vive.com/us/product/vive-xr-elite/overview/ HTC Vive XR Elite Product Page]</ref>
    * Sony's [[PlayStation VR2]] (2023) included stereo passthrough, but kept it black-and-white, accessible via a dedicated button for quick checks.<ref name="RoadToVR_PSVR2">Road to VR – PSVR 2 Review (Ben Lang, Feb 22, 2023)</ref>
    * The [[Meta Quest 3]] (late 2023) brought high-resolution stereo color passthrough with an active depth sensor (structured light projector) to the mainstream consumer market, offering significantly improved clarity and depth accuracy over Quest 2 and Quest Pro.<ref name="UploadVR_Q3Review"/><ref name="Quest3PassthroughReview">[https://www.roadtovr.com/meta-quest-3-review-vr-mixed-reality-ar/ RoadToVR Quest 3 Review detailing passthrough improvements]</ref>
    * The [[Apple Vision Pro]] (2023 announcement, 2024 release) emphasized passthrough-based MR ("[[spatial computing]]"), using dual high-resolution color cameras, advanced processing ([[Apple R1]] chip), and a [[LiDAR]] scanner for precise depth mapping.<ref name="VisionProPassthrough">[https://www.apple.com/newsroom/2023/06/introducing-apple-vision-pro/ Apple Vision Pro Announcement]</ref><ref name="Verge_VisionPro">The Verge – Apple Vision Pro review: magic, until it’s not (Nilay Patel, June 2023)</ref>
    * Other high-end devices like the [[Pimax Crystal]] (2023) and [[Varjo XR-4]] (late 2023) continued to push resolution and fidelity.<ref name="PimaxCrystal">[https://pimax.com/crystal/ Pimax Crystal Product Page]</ref><ref name="VarjoXR4">[https://varjo.com/products/xr-4/ Varjo XR-4 Product Page]</ref>
    * Even mid-range devices began incorporating improved color passthrough and depth sensing, like the anticipated Pico 4 Ultra (2024).<ref name="Auganix_Pico4Ultra">Auganix – Pico Unveils Pico 4 Ultra… (Aug 21, 2024)</ref>
 
Passthrough has evolved from a basic safety utility to a core feature enabling sophisticated mixed reality experiences, blurring the lines between traditional VR and AR.
 
== Types of Passthrough ==
Passthrough implementations vary significantly. Key characteristics include:
 
=== Monochrome Passthrough ===
Uses black-and-white camera feeds. Common in earlier VR headsets (Oculus Rift S, Quest 1 & 2) or as a design choice (PSVR2), often leveraging existing grayscale tracking cameras.<ref name="UploadVR_Q3Review_MR">UploadVR – Quest 3 Review (mixed reality section)</ref><ref name="RoadToVR_PSVR2"/> Provides basic environmental awareness but lacks color cues and realism. Advantages include potentially better low-light sensitivity and lower processing requirements.<ref name="PCMag_passthrough"/>
 
=== Color Passthrough ===
Uses [[RGB]] color cameras for a full-color view of the real world, greatly enhancing realism and enabling use cases like reading phone screens or interacting with colored objects. First widely available consumer example was Meta Quest Pro.<ref name="MixedNews_Cambria"/> Quality varies significantly based on camera resolution, processing, and calibration (e.g., Quest 3 offers ~10x the passthrough pixels of Quest 2).<ref name="UploadVR_specs">UploadVR – Quest 3 Specs Compared To Quest 2 & Apple Vision Pro (David Heaney, Sep 27, 2023)</ref> High-quality color passthrough (e.g., Varjo XR series, Vision Pro) aims for near-photorealism.<ref name="Skarredghost_Varjo"/><ref name="VisionProPassthrough"/> Requires more powerful hardware and sophisticated software.
 
=== Monoscopic vs. Stereoscopic ===
*'''Monoscopic (2D):''' Uses a single camera view (or identical views) for both eyes (e.g., original HTC Vive, initial Pico 4 implementation<ref name="Reddit_PicoMono"/>). Lacks [[binocular disparity]], resulting in a "flat" image without true depth perception. Scale and distance can feel incorrect or uncomfortable.
*'''Stereoscopic (3D):''' Uses two distinct camera viewpoints (one per eye, or reconstructed dual views) to create a 3D effect with depth perception. Requires cameras positioned roughly at the user's [[interpupillary distance]] (IPD) and careful calibration/reprojection. Essential for comfortable MR and accurate spatial interaction. Implemented in Rift S, PSVR2 (B&W stereo), Quest Pro, Quest 3, Vision Pro, Varjo XR series, etc.<ref name="UploadVR_Q3Review"/> Achieving correct scale and geometry is key to avoiding discomfort.<ref name="UploadVR_Q3Review_MR"/>
 
=== Depth-Aware Passthrough ===
Systems that actively measure or infer the distance to real-world objects and surfaces, integrating this [[depth map]] into the passthrough experience. This enables:
*'''Accurate placement and scaling''' of virtual objects relative to the real world.
*'''Occlusion:''' Allowing virtual objects to realistically appear behind real objects (and vice-versa).
*'''Improved interaction:''' Understanding the geometry of the environment for physics and hand interactions.
Methods include:
*'''Passive Stereo Vision:''' Calculating depth from the differences between two camera images (computationally intensive, can struggle with textureless surfaces).
*'''Active Depth Sensing:''' Using dedicated sensors like Infrared (IR) projectors (Quest Pro inference<ref name="MixedNews_Cambria"/>), Structured Light projectors (Quest 3<ref name="UploadVR_Q3Review"/>), [[Time-of-Flight]] (ToF) sensors, or [[LiDAR]] (Vision Pro<ref name="VisionProPassthrough"/>). These provide more robust and direct depth measurements.
Depth-aware passthrough significantly enhances MR realism and comfort, enabling features like automatic room scanning and persistent virtual object anchoring.<ref name="UploadVR_Q3Review_MR"/>
 
=== Mixed Reality Blending ===
Refers to how seamlessly the passthrough system integrates virtual content with the real-world camera feed. Advanced implementations aim to unify lighting, shadows, reflections, and occlusion across both realities. Examples include:
*Virtual objects casting realistic shadows on real surfaces.<ref name="Varjo_blog">Varjo Blog – Video Pass-Through XR – Merge Real and Virtual (Urho Konttori, 2020)</ref>
*Virtual elements being correctly hidden by real furniture or people.
*Virtual lighting affecting the appearance of the real world within the passthrough view (and vice-versa).
*Using [[Machine Learning|ML]] for scene segmentation (identifying walls, floors, furniture, people) to enable complex interactions.<ref name="XRToday_def"/>
Requires high-quality color, stereoscopic depth, active depth sensing, low latency, and sophisticated rendering techniques (e.g., real-time lighting estimation, environmental mapping). Devices like Quest 3 and Vision Pro heavily emphasize these capabilities.<ref name="UploadVR_Q3Review_MR"/><ref name="Verge_VisionPro"/>
 
== Technical Challenges ==
Creating high-quality, comfortable passthrough involves overcoming significant hurdles:
 
*'''[[Latency (engineering)|Latency]]:''' The delay between real-world motion and the passthrough display update ([[photon-to-photon latency]]). High latency (>~20ms<ref name="LatencyThreshold"/>) causes disorientation, [[motion sickness]] ("world swimming"), and breaks immersion. Fast processing pipelines are essential.<ref name="PassthroughChallengesUploadVR"/> Residual latency can cause ghosting or trailing artifacts on moving objects.<ref name="UploadVR_ghosting">UploadVR – Quest 3 Review (on ghosting issue)</ref>
*'''Resolution and Image Quality:''' Camera feeds are often lower resolution than human vision, leading to pixelation or blurriness, making fine details (like text) hard to see.<ref name="CameraLimitations">[https://arstechnica.com/gadgets/2023/10/quest-3-review-finally-real-mixed-reality-for-under-500/ Ars Technica Quest 3 Review discussing passthrough quality]</ref> Limited [[dynamic range]] struggles with bright highlights and dark shadows compared to the human eye. Poor [[low-light performance]] results in noisy, grainy images.<ref name="CameraLimitations"/> Achieving high resolution and good image quality requires better sensors and significant processing power.
*'''Camera Placement and Perspective Mismatch:''' Cameras are offset from the user's eyes, causing [[parallax]] errors if not corrected. Naive display leads to distorted views, incorrect scale, and depth perception issues, especially for close objects.<ref name="PassthroughChallengesUploadVR"/> Sophisticated [[reprojection]] algorithms are needed to warp the camera view to match the eye's perspective, but perfect correction is difficult.<ref name="KGuttag_Align"/> This geometric misalignment can cause eye strain or discomfort.<ref name="KGuttag_Align"/> Close objects (<~0.5m) often appear warped even in good systems due to sensor/lens limitations and reprojection challenges.<ref name="UploadVR_Q3Review_MR"/>
*'''Depth Perception and Occlusion:''' Even with stereo cameras, accurately replicating human depth perception is hard. Incorrect [[IPD]] matching or calibration can lead to scale issues. Lack of accurate, real-time depth maps makes correct occlusion (virtual behind real) difficult, breaking immersion. Errors in depth sensing or fusion can cause virtual objects to flicker or appear incorrectly positioned.<ref name="StereoPassthrough"/>
*'''Color Accuracy and Calibration:''' Matching the colors and brightness of the passthrough feed to both the real world and virtual elements is challenging. Poor white balance or color calibration makes the view look unnatural or filtered.<ref name="CameraLimitations"/> Display limitations also affect color reproduction. Consistent calibration across cameras and over time (accounting for thermal drift) is crucial.<ref name="KGuttag_Align"/>
*'''[[Field of View]] (FOV):''' Passthrough FOV is often narrower than human vision or even the headset's display FOV, creating a "tunnel vision" effect or visible borders where the passthrough image ends. Wide-angle lenses used to increase FOV introduce distortion that needs correction.
 
== Modern Solutions and Advancements ==
Engineers employ various techniques to address passthrough challenges:
 
*'''Multi-Camera Sensor Fusion:''' Using multiple cameras with different strengths (e.g., high-resolution RGB for color, fast monochrome for low-latency motion/detail) and fusing their data computationally.<ref name="MixedNews_Cambria"/> Overlapping camera views help compute stereo depth and increase effective FOV.<ref name="XRToday_def"/>
*'''Active Depth Sensing:''' Incorporating dedicated depth sensors (IR ToF, Structured Light, [[LiDAR]]) provides robust, real-time 3D geometry information of the environment, improving reprojection accuracy, occlusion handling, and spatial anchoring.<ref name="UploadVR_Q3Review"/><ref name="VisionProPassthrough"/> This enables features like quick room meshing via APIs (e.g., Meta's Spatial Anchors, Apple's ARKit/RoomPlan).
*'''[[Machine Learning]] Enhancements:''' Using AI/ML for various tasks:
    * Image upscaling and denoising to improve clarity, especially in low light.
    * Advanced [[reprojection]] algorithms for more accurate perspective correction.<ref name="QuestProPassthrough"/>
    * Scene segmentation to identify objects (hands, people, furniture) for better interaction and occlusion.<ref name="XRToday_def"/>
    * Improving [[Simultaneous localization and mapping|SLAM]] for more stable tracking and anchoring of virtual objects.
*'''Reprojection and Virtual Cameras:''' Software techniques that warp the captured camera images based on depth data to synthesize a view from the user's actual eye positions ("virtual cameras"<ref name="KGuttag_Align"/>). [[Asynchronous TimeWarp|Time-warping]] techniques can further reduce perceived latency by adjusting the image based on last-moment head movements.
*'''Improved Optics and Displays:''' [[Pancake lens|Pancake lenses]] allow for thinner headsets where cameras can potentially be placed closer to the eyes, reducing offset. Higher resolution, higher [[dynamic range]] (e.g., [[Micro-OLED]] in Vision Pro), and faster refresh rate displays improve the fidelity of the displayed passthrough feed. Careful calibration of lens distortion profiles is also applied.<ref name="RoadToVR_PSVR2"/>
*'''User Experience (UX) Improvements:''' Features like a dedicated passthrough toggle button (PSVR2<ref name="RoadToVR_PSVR2"/>), automatic passthrough activation when nearing boundaries (Quest Guardian<ref name="UploadVR_Q3Review_MR"/>), and boundaryless MR modes enhance usability and seamlessly blend real/virtual interactions.
 
== Applications and Use Cases ==
Passthrough enables diverse applications by allowing users to interact with the real world while immersed:
 
=== Consumer Uses ===
*'''Safety and Convenience:''' Defining play boundaries ([[Guardian system]], [[Chaperone (virtual reality)]]), avoiding obstacles, checking phones, finding controllers, or interacting briefly with people/pets without removing the headset.<ref name="PCMag_passthrough"/>
*'''[[Mixed Reality]] Gaming and Entertainment:''' Games where virtual elements interact with the user's physical room (e.g., characters hiding behind real furniture, virtual objects placed on real tables).<ref name="UploadVR_Q3Review_MR"/> Creative apps allowing virtual painting on real walls.
*'''Productivity and Utility:''' Using [[virtual desktop]]s or multiple virtual monitors while still seeing the real keyboard, mouse, and desk.<ref name="TechTarget_ARdef">[https://www.techtarget.com/whatis/definition/augmented-reality-AR TechTarget: What is augmented reality (AR)?]</ref>
*'''Social Presence:''' Reducing isolation during VR use by allowing users to see others in the same physical space. Enabling co-located MR experiences where multiple users interact with shared virtual content in the same room.
 
=== Enterprise and Professional Uses ===
*'''Collaboration:''' Design reviews where virtual prototypes are viewed in a real meeting room alongside physical mockups or colleagues.<ref name="XRToday_enterprise">XR Today – VR Passthrough in Enterprise (Immersive Learning News)</ref> Remote collaboration where experts guide on-site technicians using virtual annotations overlaid on the real equipment view.
*'''Training and Simulation:''' Combining virtual scenarios with physical controls or environments (e.g., flight simulation using a real cockpit visible via passthrough, medical training on physical manikins with virtual overlays).<ref name="VIVE_Blog_Sauce"/>
*'''Visualization:''' Architects visualizing 3D models on a real site, designers overlaying virtual concepts onto physical products.
*'''Productivity:''' Creating expansive virtual workspaces integrated with the physical office environment, improving multitasking while maintaining awareness.<ref name="XRToday_benefits">XR Today – What is VR Passthrough... (on benefits of passthrough)</ref>
 
=== Industrial and Field Uses ===
*'''Maintenance and Repair:''' Displaying step-by-step instructions, diagrams, or real-time data directly overlaid onto the machinery being worked on.
*'''Assembly and Manufacturing:''' Providing guidance and quality control checks by highlighting parts or showing virtual indicators on physical products.
*'''Logistics:''' Warehouse workers seeing picking information or navigation paths overlaid onto the real warehouse environment.
*'''Construction:''' On-site visualization of BIM models overlaid onto the actual construction progress for inspection and alignment checks.
*'''Remote Operation:''' Controlling robots or drones using a passthrough view from the machine's perspective, augmented with virtual data displays.
 
== Comparison with Optical See-Through AR ==
Passthrough (Video See-Through, VST) is distinct from [[Optical See-Through]] (OST) AR, used by devices like [[Microsoft HoloLens|HoloLens]] and [[Magic Leap]].
 
*'''Optical See-Through (OST):'''
    * Uses semi-transparent displays ([[waveguide]]s, birdbath optics) allowing direct view of the real world. Virtual images are projected onto these combiners.
    * '''Pros:''' Real world is seen perfectly naturally (zero latency, full resolution/color/dynamic range). Lower power consumption for viewing the real world. Solid real-world objects are always solid.
    * '''Cons:''' Virtual elements often appear transparent or "ghostly," lacking solidity. Limited [[Field of View|FOV]] for virtual content is common. Difficulty displaying black (virtual content is additive). Virtual content can be washed out by bright ambient light. Accurate alignment ("registration") of virtual content to the real world can be challenging. Cannot computationally modify the real-world view.
*'''Video Passthrough (VST):'''
    * Uses cameras and opaque displays to show a reconstructed view of the real world.
    * '''Pros:''' Virtual elements can be fully opaque and seamlessly blended. Potential for wider FOV matching the VR display. Can computationally modify the real-world view (e.g., brightness enhancement, selective filtering). Better blocking of ambient light for virtual content.
    * '''Cons:''' Real-world view is mediated by technology, subject to limitations (latency, resolution, color, dynamic range, distortion). Higher power consumption. Potential for discomfort (motion sickness, eye strain) if not implemented well. Real-world objects might appear less "solid" due to latency or artifacts.<ref name="SkarbezVSTvsOST"/>
 
VST AR is currently favored in the consumer MR space, leveraging existing VR display technology, while OST AR maintains advantages for applications where unobstructed real-world vision is paramount.
 
== Notable Implementations ==
*'''[[Meta Quest]] Series (Quest, Quest 2, Quest Pro, Quest 3):''' Evolved from basic monochrome safety features to sophisticated, depth-aware color passthrough using [[Machine Learning|ML]] reconstruction, making MR central to the platform.<ref name="Quest3PassthroughReview"/><ref name="QuestProPassthrough"/>
*'''[[Apple Vision Pro]]:''' High-resolution color passthrough as the default mode for "[[Spatial Computing]]", emphasizing low latency via a dedicated [[Apple R1]] chip and [[LiDAR]] for depth.<ref name="VisionProPassthrough"/>
*'''[[Varjo]] XR Series (XR-1, XR-3, XR-4):''' Industry benchmark for high-fidelity, low-latency color passthrough, aimed at professional/enterprise markets.<ref name="VarjoXR4"/><ref name="Skarredghost_Varjo"/>
*'''[[HTC Vive XR Elite]]:''' Offers color passthrough with a depth sensor for MR capabilities.<ref name="ViveXRElite"/>
*'''[[Pimax Crystal]]:''' High-resolution VR headset incorporating color passthrough features.<ref name="PimaxCrystal"/>
*'''[[Lynx R1]]:''' Standalone headset project focusing specifically on delivering quality color passthrough at a competitive price point.<ref name="LynxR1">[https://www.lynx-r.com/ Lynx R1 Official Website]</ref>
*'''[[PlayStation VR2]]:''' Features stereo black-and-white passthrough primarily for setup and quick environment checks.<ref name="RoadToVR_PSVR2"/>
*'''[[Valve Index]]:''' Basic stereoscopic monochrome passthrough via front cameras.<ref name="Index_Docs">Valve Index Hardware Documentation. (2019).</ref>
 
== Future Developments ==
Ongoing research and development aim to improve passthrough by:
*  Achieving even lower latency and higher resolution/FOV, approaching the fidelity of human vision.
*  Improving camera [[dynamic range]], color fidelity, and low-light performance.
*  Developing more sophisticated and efficient [[depth sensing]] and real-time 3D reconstruction (e.g., using [[LiDAR]], advanced [[Computer Vision|CV]], [[Neural Radiance Fields|NeRFs]]).
*  Integrating [[Artificial Intelligence|AI]] for enhanced scene understanding, object recognition, segmentation, and interaction modeling (realistic physics, occlusion).
*  Implementing selective passthrough (showing only specific real-world elements like hands or keyboards) and potentially "augmented reality" filters applied to the real-world view.
*  Utilizing [[eye tracking]] for [[foveated rendering]] of the passthrough feed or dynamic depth-of-field adjustments.
*  Exploring novel camera technologies like light field cameras (e.g., Meta's "Flamera" concept<ref name="KGuttag_Flamera">KGOnTech (Karl Guttag) – Meta Flamera Light Field Passthrough</ref>) to better solve perspective issues.
 
As technology matures, VST passthrough aims to provide a near-seamless blend between the virtual and physical worlds, potentially unifying VR and AR capabilities into single, versatile devices.
 
== See Also ==
*  [[Mixed Reality]] (MR)
*  [[Augmented Reality]] (AR)
*  [[Virtual Reality]] (VR)
*  [[Spatial computing]]
*  [[Latency (engineering)]]
*  [[Computer Vision]]
*  [[Head-mounted display]] (HMD)
*  [[Depth map]]
*  [[Simultaneous localization and mapping]] (SLAM)
*  [[Roomscale]] VR
*  [[Guardian system]] (Meta)
*  [[Chaperone (virtual reality)]] (SteamVR)
*  [[Optical See-Through]]
*  [[Microsoft HoloLens]]
*  [[Magic Leap]]
*  [[Varjo]]
*  [[Meta Quest 3]]
*  [[Apple Vision Pro]]
 
== References ==
<references>
<ref name="XRToday_def">XR Today – What is VR Passthrough and How is it Shaping the Future of XR? (Immersive Learning News, Dec 2024)</ref>
<ref name="VIVE_Blog_Sauce">[https://blog.vive.com/us/what-is-vr-passthrough-mixed-realitys-secret-sauce/ VIVE Blog: What is VR Passthrough? Mixed Reality's Secret Sauce]</ref>
<ref name="SkarbezVSTvsOST">[https://www.researchgate.net/publication/315722770_Revisiting_Milgram_and_Kishino%27s_Reality-Virtuality_Continuum Revisiting Milgram and Kishino's Reality-Virtuality Continuum] - Discusses the spectrum including Video See-Through.</ref>
<ref name="StereoPassthrough">[https://ieeexplore.ieee.org/document/9191148 Example paper discussing stereoscopic passthrough challenges]</ref>
<ref name="MixedNews_Cambria">MIXED News – Project Cambria: Meta explains new passthrough technology (Tomislav Bezmalinović, May 16, 2022)</ref>
<ref name="PassthroughChallengesUploadVR">[https://uploadvr.com/passthrough-ar-technical-challenges/ Passthrough AR: The Technical Challenges of Blending Realities] - UploadVR article discussing latency, distortion, etc.</ref>
<ref name="KGuttag_Align">KGOnTech (Karl Guttag) – Perspective Correct Passthrough (Sept 26, 2023)</ref>
<ref name="QuestProPassthrough">[https://www.meta.com/blog/quest/meta-reality-passthrough-quest-pro/ Meta Blog: Inside Meta Reality and Passthrough on Quest Pro]</ref>
<ref name="UploadVR_Q3Review">UploadVR – Quest 3 Review: Excellent VR With Limited MR (David Heaney, Oct 9, 2023)</ref>
<ref name="LatencyThreshold">[https://research.nvidia.com/publication/2016-07_Latency-Requirements-Plausible-Interaction-Augmented-and-Virtual-Reality Latency Requirements for Plausible Interaction in Augmented and Virtual Reality] - Research discussing latency impact.</ref>
<ref name="MilgramKishino1994">Milgram, P., & Kishino, F. (1994). A taxonomy of mixed reality visual displays. IEICE Transactions on Information Systems, E77-D(12), 1321-1329.</ref>
<ref name="RoadToVR_Vive">Road to VR – 8 Minutes of the HTC Vive’s Front-facing Camera in Action (Paul James, Mar 10, 2016)</ref>
<ref name="RiftS_Docs">Oculus Rift S Product Documentation. (2019).</ref>
<ref name="Quest2Passthrough">[https://www.meta.com/blog/quest/oculus-quest-2-passthrough-public-api-passthrough/ Quest Blog on Quest 2 Passthrough improvements]</ref>
<ref name="PCMag_passthrough">PCMag – Oculus Experiments With Mixed Reality via New Passthrough API (Nathaniel Mott, July 25, 2021)</ref>
<ref name="Skarredghost_Varjo">The Ghost Howls – Varjo XR-3 hands-on review (Tony Vitillo, June 8, 2022)</ref>
<ref name="Reddit_PicoMono">Reddit – Meta Quest Pro vs PICO 4 Passthrough Comparison (discussion summary, 2022)</ref>
<ref name="ViveXRElite">[https://www.vive.com/us/product/vive-xr-elite/overview/ HTC Vive XR Elite Product Page]</ref>
<ref name="RoadToVR_PSVR2">Road to VR – PSVR 2 Review (Ben Lang, Feb 22, 2023)</ref>
<ref name="Quest3PassthroughReview">[https://www.roadtovr.com/meta-quest-3-review-vr-mixed-reality-ar/ RoadToVR Quest 3 Review detailing passthrough improvements]</ref>
<ref name="VisionProPassthrough">[https://www.apple.com/newsroom/2023/06/introducing-apple-vision-pro/ Apple Vision Pro Announcement]</ref>
<ref name="Verge_VisionPro">The Verge – Apple Vision Pro review: magic, until it’s not (Nilay Patel, June 2023)</ref>
<ref name="PimaxCrystal">[https://pimax.com/crystal/ Pimax Crystal Product Page]</ref>
<ref name="VarjoXR4">[https://varjo.com/products/xr-4/ Varjo XR-4 Product Page]</ref>
<ref name="Auganix_Pico4Ultra">Auganix – Pico Unveils Pico 4 Ultra… (Aug 21, 2024)</ref>
<ref name="UploadVR_Q3Review_MR">UploadVR – Quest 3 Review (mixed reality section)</ref>
<ref name="UploadVR_specs">UploadVR – Quest 3 Specs Compared To Quest 2 & Apple Vision Pro (David Heaney, Sep 27, 2023)</ref>
<ref name="UploadVR_ghosting">UploadVR – Quest 3 Review (on ghosting issue)</ref>
<ref name="CameraLimitations">[https://arstechnica.com/gadgets/2023/10/quest-3-review-finally-real-mixed-reality-for-under-500/ Ars Technica Quest 3 Review discussing passthrough quality]</ref>
<ref name="Varjo_blog">Varjo Blog – Video Pass-Through XR – Merge Real and Virtual (Urho Konttori, 2020)</ref>
<ref name="TechTarget_ARdef">[https://www.techtarget.com/whatis/definition/augmented-reality-AR TechTarget: What is augmented reality (AR)?]</ref>
<ref name="XRToday_enterprise">XR Today – VR Passthrough in Enterprise (Immersive Learning News)</ref>
<ref name="XRToday_benefits">XR Today – What is VR Passthrough... (on benefits of passthrough)</ref>
<ref name="Index_Docs">Valve Index Hardware Documentation. (2019).</ref>
<ref name="LynxR1">[https://www.lynx-r.com/ Lynx R1 Official Website]</ref>
<ref name="KGuttag_Flamera">KGOnTech (Karl Guttag) – Meta Flamera Light Field Passthrough</ref>
<!-- Citations from Article 4 (if distinct and useful, integrate above or add here) -->
<ref name="Bastug2017">Bastug, E., Bennis, M., Médard, M., & Debbah, M. (2017). Toward Interconnected Virtual Reality: Opportunities, Challenges, and Enablers. IEEE Communications Magazine, 55(6), 110-117.</ref> <!-- General VR context -->
<ref name="MetaConnect2022_QuestPro">Meta Connect 2022: Technical Overview of Quest Pro's Passthrough Technology. (2022).</ref> <!-- Covered by QuestProPassthrough ref -->
<ref name="AppleVisionPro_Specs">Apple Vision Pro Technical Specifications. (2024).</ref> <!-- Covered by VisionProPassthrough ref -->
<ref name="Quest_ProductDocs2023">Meta Quest Product Documentation. (2023).</ref> <!-- Generic, specific reviews are better -->
<ref name="ViveXRElite_Specs">HTC Vive XR Elite Product Specifications. (2023).</ref> <!-- Covered by ViveXRElite ref -->
<ref name="AppleVisionPro_TechOverview">Apple Vision Pro Technical Overview. (2024).</ref> <!-- Covered by VisionProPassthrough ref -->
<ref name="Pico4_Docs">Pico 4 Product Documentation. (2022).</ref> <!-- Generic -->
<ref name="McGill2022_Comparison">McGill, M., Boland, D., Zielinski, R., & Brewster, S. (2022). Comparing consumer virtual reality headset passthrough, augmented reality head-mounted displays, and augmented virtuality: evaluating performance, presence, and user experience. Virtual Reality, 26, 267-289.</ref> <!-- Good academic comparison -->
<ref name="Steinicke2010_Presence">Steinicke, F., Bruder, G., Hinrichs, K., & Steed, A. (2010). Presence and performance in virtual environments: concepts and measures. IEEE Transactions on Visualization and Computer Graphics, 16(5), 645-652.</ref> <!-- General presence/latency -->
<ref name="Seetharaman2020_EvalVST">Seetharaman, V., Lee, S., & Yang, Y. (2020). Evaluation of video see-through techniques for mixed reality applications. IEEE Access, 8, 156057-156066.</ref> <!-- Academic VST evaluation -->
<ref name="Hu2023_FOV_Eval">Hu, C., Li, D., Wu, C. et al. (2023). Quantitative evaluation of the field of view for current VR/AR passthrough systems. Scientific Reports, 13, 11574.</ref> <!-- FOV specific research -->
<ref name="Park2022_AdapTutAR">Park, S., Gebhardt, C., Rädle, R., Feit, A. M., Vrzakova, H., Dayama, N. R., Yeo, H., Klokmose, C. N., Quigley, A., Oulasvirta, A., & Hilliges, O. (2022). AdapTutAR: An adaptive tutoring system for machine tasks in augmented reality. IEEE Transactions on Visualization and Computer Graphics, 28(4), 1819-1830.</ref> <!-- Specific AR application -->
<ref name="Cheng2021_MutualActuation">Cheng, L., Marwecki, S., & Baudisch, P. (2021). Mutual human actuation: Enabling safety features for walk-around VR. In IEEE Virtual Reality and 3D User Interfaces (VR), pp. 361-370.</ref> <!-- Safety research -->
<ref name="Shen2021_VisionHandAR">Shen, Y., Ong, S. K., & Nee, A. Y. (2021). Vision-based hand interaction in augmented reality environment. International Journal of Human-Computer Interaction, 37(1), 54-77.</ref> <!-- Hand interaction research -->
<ref name="Kim2019_ARSmartGlasses">Kim, S., Nussbaum, M. A., & Gabbard, J. L. (2019). Augmented reality "smart glasses" in the workplace: Industry perspectives and challenges for worker safety and health. IIE Transactions on Occupational Ergonomics and Human Factors, 7(3-4), 253-258.</ref> <!-- Workplace AR context -->
<ref name="MetaResearchBlog_MRAdvances">Meta Research Blog: Advances in Mixed Reality Technology. (2024).</ref> <!-- General industry progress -->
<ref name="Anthes2023_StateOfArtVR">Anthes, C., García-Hernández, R. J., Wiedemann, M., & Kranzlmüller, D. (2023). State of the art of virtual reality technologies. In IEEE Aerospace Conference, pp. 1-19.</ref> <!-- VR state of art -->
<ref name="Chan2024_NextGenMR">Chan, H., Zou, J., Blackmon, T., & Yang, W. (2024). The next generation of mixed reality: Challenges and opportunities. IEEE Transactions on Visualization and Computer Graphics, 30(2), 812-831.</ref> <!-- Future MR research -->
<!-- Citations from Article 3 (if distinct and useful) -->
<ref name="UploadVR_Q3PassthroughUpdate">[https://www.uploadvr.com/quest-v64-update-improved-passthrough/ UploadVR: Quest 3 Gets Improved Passthrough Quality & New Features]</ref> <!-- Covered by Q3 review refs -->
<ref name="KGonTech_QuestProCritique">[https://kguttag.com/2023/01/03/meta-quest-pro-part-1-unbelievably-bad-ar-passthrough/ KGonTech: Meta Quest Pro Passthrough Analysis]</ref> <!-- Covered by KGuttag_Align ref potentially, or good as specific critique -->
<ref name="Wiki_Quest3">Wikipedia: Meta Quest 3</ref> <!-- Wikipedia shouldn't cite itself generally -->
<ref name="VRExpert_WhatIsPassthrough">[https://vrx.vr-expert.com/what-is-pass-through-and-why-do-all-vr-headsets-have-it/ VR Expert: What is pass-through, and why do all VR headsets have it?]</ref> <!-- General overview -->
<ref name="Mashable_Q3HandsOn">[https://mashable.com/article/meta-quest-3-hands-on Mashable: Meta Quest 3 hands-on review]</ref> <!-- Covered by other Q3 reviews -->
<ref name="MetaNews_QuestProLaunch">[https://about.fb.com/news/2022/10/meta-quest-pro-is-now-available/ Meta News: Meta Quest Pro is Now Available]</ref> <!-- Launch info -->
<ref name="VRExpert_ActivateQ3Passthrough">[https://knowledge.vr-expert.com/kb/how-to-activate-passthrough-on-the-meta-quest-3/ VR Expert: How To Activate Passthrough On The Meta Quest 3]</ref> <!-- Specific UI instruction -->
<ref name="ResolutionGames_PassthroughPromise">[https://www.resolutiongames.com/blog/the-power-and-promise-of-passthrough Resolution Games: The Power and Promise of Passthrough]</ref> <!-- Developer perspective -->
</references>

Revision as of 07:17, 25 April 2025

Template:Featured Passthrough, often referred to as video passthrough, is a feature found in Virtual Reality (VR) and Mixed Reality (MR) headsets that utilizes external cameras to capture a live video feed of the physical environment around the user and display it on the internal screens within the headset.[1][2] This capability effectively allows users to see the "real world" without removing the headset, bridging the gap between fully immersive virtual experiences and the user's actual surroundings.

While primarily a feature of VR headsets aiming to add environmental awareness or MR capabilities, it functions as a form of Augmented Reality (AR), often termed "Video See-Through AR" (VST AR) or sometimes "pseudo-AR," as opposed to "Optical See-Through AR" (OST AR) systems which use transparent displays.[3] Passthrough is a key enabler of mixed reality and spatial computing experiences on modern headsets.

Core Technology and How It Works

The fundamental principle of passthrough involves a real-time processing pipeline:

  1. Capture: One or more outward-facing digital cameras mounted on the headset capture video of the external world. Early or basic systems might use a single camera (providing a monoscopic view), while more advanced systems use two or more cameras to capture stereoscopic video, enabling depth perception.[4] Modern systems often use a combination of RGB color cameras and monochrome (grayscale) sensors for different purposes (e.g., capturing color data vs. motion/detail).[5]
  2. Processing: The captured video footage is sent to the headset's processor (either an onboard SoC or a connected PC's GPU). This stage is computationally intensive and critical for a usable and comfortable experience. It typically involves several steps:
    • Rectification/Undistortion: Correcting lens distortion inherent in the wide-angle cameras typically used to maximize FOV.
    • Reprojection/Warping: Adjusting the captured image perspective to align with the user's eye position inside the headset, rather than the camera's physical position on the outside. This difference in viewpoint causes parallax, and correcting it ("perspective correction") is crucial for accurate spatial representation, correct scale perception, and minimizing motion sickness.[6][7] Algorithms based on Computer Vision and potentially IMU sensor data are used. Some modern headsets, like the Meta Quest Pro and Meta Quest 3, employ Machine Learning or neural networks to improve the realism and accuracy of this reconstruction.[8]
    • Sensor Fusion: Combining data from multiple cameras (e.g., fusing monochrome detail with RGB color[5]) and integrating tracking data (e.g., from inside-out tracking sensors or depth sensors) to ensure the passthrough view remains stable, depth-correct, and aligned with the user's head movements.
    • Color Correction & Enhancement: Adjusting colors, brightness, and contrast to appear more natural, especially under varying lighting conditions. This can also involve AI-based denoising or upscaling.[9]
  3. Display: The processed video feed is rendered onto the headset's internal displays, replacing or being overlaid upon the virtual content. The primary goal is to achieve this entire pipeline with minimal latency (ideally under 20 milliseconds[10]) to avoid discomfort and maintain realism.

History and Evolution

While the concept of video passthrough existed in research labs for decades,[11] its implementation in consumer VR headsets evolved significantly:

  • Early Stages (Mid-2010s): Passthrough began appearing primarily as a safety feature. In 2016, the HTC Vive prototype (Vive Pre) introduced a front-facing camera providing a basic, monochrome, 2D view for obstacle avoidance. Valve's software projected this onto a virtual sphere to approximate perspective.[12] It was low-resolution and intended for brief checks.
  • Integrated Monochrome (Late 2010s): Headsets using inside-out tracking leveraged their tracking cameras for improved passthrough. The Oculus Rift S (2019) offered "Passthrough+" using its multiple monochrome cameras for a stereoscopic view.[13] The original Oculus Quest (2019) and Meta Quest 2 (2020) provided similar basic monochrome passthrough, mainly for setting up the Guardian system and quick environment checks.[14]
  • Early Mixed Reality Steps (Early 2020s): In 2021, Meta released an experimental Passthrough API for Quest 2 developers, allowing apps to overlay virtual elements onto the monochrome feed, marking a step towards consumer MR.[15] Simultaneously, enterprise headsets like the Varjo XR-1 (2019) and XR-3 (2021) pushed high-fidelity color passthrough with dual high-resolution cameras, setting a benchmark for quality.[16]
  • Mainstream Color Passthrough (2022-Present):
   * The Meta Quest Pro (2022) was the first major consumer headset featuring high-quality, stereoscopic color passthrough, using a novel camera array (monochrome for depth/detail, RGB for color) and ML reconstruction.[5]
   * Competitors like the Pico 4 (late 2022) and HTC Vive XR Elite (2023) also introduced color passthrough, although early implementations like the Pico 4's were initially monoscopic and lacked depth correction.[17][18]
   * Sony's PlayStation VR2 (2023) included stereo passthrough, but kept it black-and-white, accessible via a dedicated button for quick checks.[19]
   * The Meta Quest 3 (late 2023) brought high-resolution stereo color passthrough with an active depth sensor (structured light projector) to the mainstream consumer market, offering significantly improved clarity and depth accuracy over Quest 2 and Quest Pro.[9][20]
   * The Apple Vision Pro (2023 announcement, 2024 release) emphasized passthrough-based MR ("spatial computing"), using dual high-resolution color cameras, advanced processing (Apple R1 chip), and a LiDAR scanner for precise depth mapping.[21][22]
   * Other high-end devices like the Pimax Crystal (2023) and Varjo XR-4 (late 2023) continued to push resolution and fidelity.[23][24]
   * Even mid-range devices began incorporating improved color passthrough and depth sensing, like the anticipated Pico 4 Ultra (2024).[25]

Passthrough has evolved from a basic safety utility to a core feature enabling sophisticated mixed reality experiences, blurring the lines between traditional VR and AR.

Types of Passthrough

Passthrough implementations vary significantly. Key characteristics include:

Monochrome Passthrough

Uses black-and-white camera feeds. Common in earlier VR headsets (Oculus Rift S, Quest 1 & 2) or as a design choice (PSVR2), often leveraging existing grayscale tracking cameras.[26][19] Provides basic environmental awareness but lacks color cues and realism. Advantages include potentially better low-light sensitivity and lower processing requirements.[15]

Color Passthrough

Uses RGB color cameras for a full-color view of the real world, greatly enhancing realism and enabling use cases like reading phone screens or interacting with colored objects. First widely available consumer example was Meta Quest Pro.[5] Quality varies significantly based on camera resolution, processing, and calibration (e.g., Quest 3 offers ~10x the passthrough pixels of Quest 2).[27] High-quality color passthrough (e.g., Varjo XR series, Vision Pro) aims for near-photorealism.[16][21] Requires more powerful hardware and sophisticated software.

Monoscopic vs. Stereoscopic

  • Monoscopic (2D): Uses a single camera view (or identical views) for both eyes (e.g., original HTC Vive, initial Pico 4 implementation[17]). Lacks binocular disparity, resulting in a "flat" image without true depth perception. Scale and distance can feel incorrect or uncomfortable.
  • Stereoscopic (3D): Uses two distinct camera viewpoints (one per eye, or reconstructed dual views) to create a 3D effect with depth perception. Requires cameras positioned roughly at the user's interpupillary distance (IPD) and careful calibration/reprojection. Essential for comfortable MR and accurate spatial interaction. Implemented in Rift S, PSVR2 (B&W stereo), Quest Pro, Quest 3, Vision Pro, Varjo XR series, etc.[9] Achieving correct scale and geometry is key to avoiding discomfort.[26]

Depth-Aware Passthrough

Systems that actively measure or infer the distance to real-world objects and surfaces, integrating this depth map into the passthrough experience. This enables:

  • Accurate placement and scaling of virtual objects relative to the real world.
  • Occlusion: Allowing virtual objects to realistically appear behind real objects (and vice-versa).
  • Improved interaction: Understanding the geometry of the environment for physics and hand interactions.

Methods include:

  • Passive Stereo Vision: Calculating depth from the differences between two camera images (computationally intensive, can struggle with textureless surfaces).
  • Active Depth Sensing: Using dedicated sensors like Infrared (IR) projectors (Quest Pro inference[5]), Structured Light projectors (Quest 3[9]), Time-of-Flight (ToF) sensors, or LiDAR (Vision Pro[21]). These provide more robust and direct depth measurements.

Depth-aware passthrough significantly enhances MR realism and comfort, enabling features like automatic room scanning and persistent virtual object anchoring.[26]

Mixed Reality Blending

Refers to how seamlessly the passthrough system integrates virtual content with the real-world camera feed. Advanced implementations aim to unify lighting, shadows, reflections, and occlusion across both realities. Examples include:

  • Virtual objects casting realistic shadows on real surfaces.[28]
  • Virtual elements being correctly hidden by real furniture or people.
  • Virtual lighting affecting the appearance of the real world within the passthrough view (and vice-versa).
  • Using ML for scene segmentation (identifying walls, floors, furniture, people) to enable complex interactions.[1]

Requires high-quality color, stereoscopic depth, active depth sensing, low latency, and sophisticated rendering techniques (e.g., real-time lighting estimation, environmental mapping). Devices like Quest 3 and Vision Pro heavily emphasize these capabilities.[26][22]

Technical Challenges

Creating high-quality, comfortable passthrough involves overcoming significant hurdles:

  • Latency: The delay between real-world motion and the passthrough display update (photon-to-photon latency). High latency (>~20ms[10]) causes disorientation, motion sickness ("world swimming"), and breaks immersion. Fast processing pipelines are essential.[6] Residual latency can cause ghosting or trailing artifacts on moving objects.[29]
  • Resolution and Image Quality: Camera feeds are often lower resolution than human vision, leading to pixelation or blurriness, making fine details (like text) hard to see.[30] Limited dynamic range struggles with bright highlights and dark shadows compared to the human eye. Poor low-light performance results in noisy, grainy images.[30] Achieving high resolution and good image quality requires better sensors and significant processing power.
  • Camera Placement and Perspective Mismatch: Cameras are offset from the user's eyes, causing parallax errors if not corrected. Naive display leads to distorted views, incorrect scale, and depth perception issues, especially for close objects.[6] Sophisticated reprojection algorithms are needed to warp the camera view to match the eye's perspective, but perfect correction is difficult.[7] This geometric misalignment can cause eye strain or discomfort.[7] Close objects (<~0.5m) often appear warped even in good systems due to sensor/lens limitations and reprojection challenges.[26]
  • Depth Perception and Occlusion: Even with stereo cameras, accurately replicating human depth perception is hard. Incorrect IPD matching or calibration can lead to scale issues. Lack of accurate, real-time depth maps makes correct occlusion (virtual behind real) difficult, breaking immersion. Errors in depth sensing or fusion can cause virtual objects to flicker or appear incorrectly positioned.[4]
  • Color Accuracy and Calibration: Matching the colors and brightness of the passthrough feed to both the real world and virtual elements is challenging. Poor white balance or color calibration makes the view look unnatural or filtered.[30] Display limitations also affect color reproduction. Consistent calibration across cameras and over time (accounting for thermal drift) is crucial.[7]
  • Field of View (FOV): Passthrough FOV is often narrower than human vision or even the headset's display FOV, creating a "tunnel vision" effect or visible borders where the passthrough image ends. Wide-angle lenses used to increase FOV introduce distortion that needs correction.

Modern Solutions and Advancements

Engineers employ various techniques to address passthrough challenges:

  • Multi-Camera Sensor Fusion: Using multiple cameras with different strengths (e.g., high-resolution RGB for color, fast monochrome for low-latency motion/detail) and fusing their data computationally.[5] Overlapping camera views help compute stereo depth and increase effective FOV.[1]
  • Active Depth Sensing: Incorporating dedicated depth sensors (IR ToF, Structured Light, LiDAR) provides robust, real-time 3D geometry information of the environment, improving reprojection accuracy, occlusion handling, and spatial anchoring.[9][21] This enables features like quick room meshing via APIs (e.g., Meta's Spatial Anchors, Apple's ARKit/RoomPlan).
  • Machine Learning Enhancements: Using AI/ML for various tasks:
   * Image upscaling and denoising to improve clarity, especially in low light.
   * Advanced reprojection algorithms for more accurate perspective correction.[8]
   * Scene segmentation to identify objects (hands, people, furniture) for better interaction and occlusion.[1]
   * Improving SLAM for more stable tracking and anchoring of virtual objects.
  • Reprojection and Virtual Cameras: Software techniques that warp the captured camera images based on depth data to synthesize a view from the user's actual eye positions ("virtual cameras"[7]). Time-warping techniques can further reduce perceived latency by adjusting the image based on last-moment head movements.
  • Improved Optics and Displays: Pancake lenses allow for thinner headsets where cameras can potentially be placed closer to the eyes, reducing offset. Higher resolution, higher dynamic range (e.g., Micro-OLED in Vision Pro), and faster refresh rate displays improve the fidelity of the displayed passthrough feed. Careful calibration of lens distortion profiles is also applied.[19]
  • User Experience (UX) Improvements: Features like a dedicated passthrough toggle button (PSVR2[19]), automatic passthrough activation when nearing boundaries (Quest Guardian[26]), and boundaryless MR modes enhance usability and seamlessly blend real/virtual interactions.

Applications and Use Cases

Passthrough enables diverse applications by allowing users to interact with the real world while immersed:

Consumer Uses

  • Safety and Convenience: Defining play boundaries (Guardian system, Chaperone (virtual reality)), avoiding obstacles, checking phones, finding controllers, or interacting briefly with people/pets without removing the headset.[15]
  • Mixed Reality Gaming and Entertainment: Games where virtual elements interact with the user's physical room (e.g., characters hiding behind real furniture, virtual objects placed on real tables).[26] Creative apps allowing virtual painting on real walls.
  • Productivity and Utility: Using virtual desktops or multiple virtual monitors while still seeing the real keyboard, mouse, and desk.[31]
  • Social Presence: Reducing isolation during VR use by allowing users to see others in the same physical space. Enabling co-located MR experiences where multiple users interact with shared virtual content in the same room.

Enterprise and Professional Uses

  • Collaboration: Design reviews where virtual prototypes are viewed in a real meeting room alongside physical mockups or colleagues.[32] Remote collaboration where experts guide on-site technicians using virtual annotations overlaid on the real equipment view.
  • Training and Simulation: Combining virtual scenarios with physical controls or environments (e.g., flight simulation using a real cockpit visible via passthrough, medical training on physical manikins with virtual overlays).[2]
  • Visualization: Architects visualizing 3D models on a real site, designers overlaying virtual concepts onto physical products.
  • Productivity: Creating expansive virtual workspaces integrated with the physical office environment, improving multitasking while maintaining awareness.[33]

Industrial and Field Uses

  • Maintenance and Repair: Displaying step-by-step instructions, diagrams, or real-time data directly overlaid onto the machinery being worked on.
  • Assembly and Manufacturing: Providing guidance and quality control checks by highlighting parts or showing virtual indicators on physical products.
  • Logistics: Warehouse workers seeing picking information or navigation paths overlaid onto the real warehouse environment.
  • Construction: On-site visualization of BIM models overlaid onto the actual construction progress for inspection and alignment checks.
  • Remote Operation: Controlling robots or drones using a passthrough view from the machine's perspective, augmented with virtual data displays.

Comparison with Optical See-Through AR

Passthrough (Video See-Through, VST) is distinct from Optical See-Through (OST) AR, used by devices like HoloLens and Magic Leap.

  • Optical See-Through (OST):
   * Uses semi-transparent displays (waveguides, birdbath optics) allowing direct view of the real world. Virtual images are projected onto these combiners.
   * Pros: Real world is seen perfectly naturally (zero latency, full resolution/color/dynamic range). Lower power consumption for viewing the real world. Solid real-world objects are always solid.
   * Cons: Virtual elements often appear transparent or "ghostly," lacking solidity. Limited FOV for virtual content is common. Difficulty displaying black (virtual content is additive). Virtual content can be washed out by bright ambient light. Accurate alignment ("registration") of virtual content to the real world can be challenging. Cannot computationally modify the real-world view.
  • Video Passthrough (VST):
   * Uses cameras and opaque displays to show a reconstructed view of the real world.
   * Pros: Virtual elements can be fully opaque and seamlessly blended. Potential for wider FOV matching the VR display. Can computationally modify the real-world view (e.g., brightness enhancement, selective filtering). Better blocking of ambient light for virtual content.
   * Cons: Real-world view is mediated by technology, subject to limitations (latency, resolution, color, dynamic range, distortion). Higher power consumption. Potential for discomfort (motion sickness, eye strain) if not implemented well. Real-world objects might appear less "solid" due to latency or artifacts.[3]

VST AR is currently favored in the consumer MR space, leveraging existing VR display technology, while OST AR maintains advantages for applications where unobstructed real-world vision is paramount.

Notable Implementations

  • Meta Quest Series (Quest, Quest 2, Quest Pro, Quest 3): Evolved from basic monochrome safety features to sophisticated, depth-aware color passthrough using ML reconstruction, making MR central to the platform.[20][8]
  • Apple Vision Pro: High-resolution color passthrough as the default mode for "Spatial Computing", emphasizing low latency via a dedicated Apple R1 chip and LiDAR for depth.[21]
  • Varjo XR Series (XR-1, XR-3, XR-4): Industry benchmark for high-fidelity, low-latency color passthrough, aimed at professional/enterprise markets.[24][16]
  • HTC Vive XR Elite: Offers color passthrough with a depth sensor for MR capabilities.[18]
  • Pimax Crystal: High-resolution VR headset incorporating color passthrough features.[23]
  • Lynx R1: Standalone headset project focusing specifically on delivering quality color passthrough at a competitive price point.[34]
  • PlayStation VR2: Features stereo black-and-white passthrough primarily for setup and quick environment checks.[19]
  • Valve Index: Basic stereoscopic monochrome passthrough via front cameras.[35]

Future Developments

Ongoing research and development aim to improve passthrough by:

  • Achieving even lower latency and higher resolution/FOV, approaching the fidelity of human vision.
  • Improving camera dynamic range, color fidelity, and low-light performance.
  • Developing more sophisticated and efficient depth sensing and real-time 3D reconstruction (e.g., using LiDAR, advanced CV, NeRFs).
  • Integrating AI for enhanced scene understanding, object recognition, segmentation, and interaction modeling (realistic physics, occlusion).
  • Implementing selective passthrough (showing only specific real-world elements like hands or keyboards) and potentially "augmented reality" filters applied to the real-world view.
  • Utilizing eye tracking for foveated rendering of the passthrough feed or dynamic depth-of-field adjustments.
  • Exploring novel camera technologies like light field cameras (e.g., Meta's "Flamera" concept[36]) to better solve perspective issues.

As technology matures, VST passthrough aims to provide a near-seamless blend between the virtual and physical worlds, potentially unifying VR and AR capabilities into single, versatile devices.

See Also

References

  1. 1.0 1.1 1.2 1.3 XR Today – What is VR Passthrough and How is it Shaping the Future of XR? (Immersive Learning News, Dec 2024)
  2. 2.0 2.1 VIVE Blog: What is VR Passthrough? Mixed Reality's Secret Sauce Cite error: Invalid <ref> tag; name "VIVE_Blog_Sauce" defined multiple times with different content
  3. 3.0 3.1 Revisiting Milgram and Kishino's Reality-Virtuality Continuum - Discusses the spectrum including Video See-Through.
  4. 4.0 4.1 Example paper discussing stereoscopic passthrough challenges
  5. 5.0 5.1 5.2 5.3 5.4 5.5 MIXED News – Project Cambria: Meta explains new passthrough technology (Tomislav Bezmalinović, May 16, 2022)
  6. 6.0 6.1 6.2 Passthrough AR: The Technical Challenges of Blending Realities - UploadVR article discussing latency, distortion, etc.
  7. 7.0 7.1 7.2 7.3 7.4 KGOnTech (Karl Guttag) – Perspective Correct Passthrough (Sept 26, 2023)
  8. 8.0 8.1 8.2 Meta Blog: Inside Meta Reality and Passthrough on Quest Pro
  9. 9.0 9.1 9.2 9.3 9.4 UploadVR – Quest 3 Review: Excellent VR With Limited MR (David Heaney, Oct 9, 2023)
  10. 10.0 10.1 Latency Requirements for Plausible Interaction in Augmented and Virtual Reality - Research discussing latency impact.
  11. Milgram, P., & Kishino, F. (1994). A taxonomy of mixed reality visual displays. IEICE Transactions on Information Systems, E77-D(12), 1321-1329.
  12. Road to VR – 8 Minutes of the HTC Vive’s Front-facing Camera in Action (Paul James, Mar 10, 2016)
  13. Oculus Rift S Product Documentation. (2019).
  14. Quest Blog on Quest 2 Passthrough improvements
  15. 15.0 15.1 15.2 PCMag – Oculus Experiments With Mixed Reality via New Passthrough API (Nathaniel Mott, July 25, 2021)
  16. 16.0 16.1 16.2 The Ghost Howls – Varjo XR-3 hands-on review (Tony Vitillo, June 8, 2022)
  17. 17.0 17.1 Reddit – Meta Quest Pro vs PICO 4 Passthrough Comparison (discussion summary, 2022)
  18. 18.0 18.1 HTC Vive XR Elite Product Page
  19. 19.0 19.1 19.2 19.3 19.4 Road to VR – PSVR 2 Review (Ben Lang, Feb 22, 2023)
  20. 20.0 20.1 RoadToVR Quest 3 Review detailing passthrough improvements
  21. 21.0 21.1 21.2 21.3 21.4 Apple Vision Pro Announcement
  22. 22.0 22.1 The Verge – Apple Vision Pro review: magic, until it’s not (Nilay Patel, June 2023)
  23. 23.0 23.1 Pimax Crystal Product Page
  24. 24.0 24.1 Varjo XR-4 Product Page
  25. Auganix – Pico Unveils Pico 4 Ultra… (Aug 21, 2024)
  26. 26.0 26.1 26.2 26.3 26.4 26.5 26.6 UploadVR – Quest 3 Review (mixed reality section)
  27. UploadVR – Quest 3 Specs Compared To Quest 2 & Apple Vision Pro (David Heaney, Sep 27, 2023)
  28. Varjo Blog – Video Pass-Through XR – Merge Real and Virtual (Urho Konttori, 2020)
  29. UploadVR – Quest 3 Review (on ghosting issue)
  30. 30.0 30.1 30.2 Ars Technica Quest 3 Review discussing passthrough quality
  31. TechTarget: What is augmented reality (AR)?
  32. XR Today – VR Passthrough in Enterprise (Immersive Learning News)
  33. XR Today – What is VR Passthrough... (on benefits of passthrough)
  34. Lynx R1 Official Website
  35. Valve Index Hardware Documentation. (2019).
  36. KGOnTech (Karl Guttag) – Meta Flamera Light Field Passthrough

Cite error: <ref> tag with name "Bastug2017" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "MetaConnect2022_QuestPro" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "AppleVisionPro_Specs" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Quest_ProductDocs2023" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "ViveXRElite_Specs" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "AppleVisionPro_TechOverview" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Pico4_Docs" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "McGill2022_Comparison" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Steinicke2010_Presence" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Seetharaman2020_EvalVST" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Hu2023_FOV_Eval" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Park2022_AdapTutAR" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Cheng2021_MutualActuation" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Shen2021_VisionHandAR" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Kim2019_ARSmartGlasses" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "MetaResearchBlog_MRAdvances" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Anthes2023_StateOfArtVR" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Chan2024_NextGenMR" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "UploadVR_Q3PassthroughUpdate" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "KGonTech_QuestProCritique" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Wiki_Quest3" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "VRExpert_WhatIsPassthrough" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "Mashable_Q3HandsOn" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "MetaNews_QuestProLaunch" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "VRExpert_ActivateQ3Passthrough" defined in <references> is not used in prior text.
Cite error: <ref> tag with name "ResolutionGames_PassthroughPromise" defined in <references> is not used in prior text.