Apple Patent | Video Pipeline

Patent: Video Pipeline

Publication Number: 20200058152

Publication Date: 20200220

Applicants: Apple

Abstract

A mixed reality system that includes a device and a base station that communicate via a wireless connection The device may include sensors that collect information about the user’s environment and about the user. The information collected by the sensors may be transmitted to the base station via the wireless connection. The base station renders frames or slices based at least in part on the sensor information received from the device, encodes the frames or slices, and transmits the compressed frames or slices to the device for decoding and display. The base station may provide more computing power than conventional stand-alone systems, and the wireless connection does not tether the device to the base station as in conventional tethered systems. The system may implement methods and apparatus to maintain a target frame rate through the wireless link and to minimize latency in frame rendering, transmittal, and display.

PRIORITY INFORMATION

[0001] This application is a continuation of PCT Application Serial No. US2018/029862, filed Apr. 27, 2018, which claims benefit of priority of U.S. Provisional Application Ser. No. 62/492,000, filed Apr. 28, 2017, the contents of which are incorporated by reference herein in their entirety.

BACKGROUND

[0002] Virtual reality (VR) systems display virtual views that provide an immersive virtual environment. Mixed reality (MR) systems combine virtual content with a view of the real world, or add virtual representations of real world objects to a virtual environment. Conventional VR and MR systems are typically either tethered systems including a base station that performs at least some of the rendering of content for display and a device connected to the base station via a physical connection (i.e., a data communications cable), or stand-alone devices that perform rendering of content locally. Stand-alone systems allow users freedom of movement; however, because of restraints including size, weight, batteries, and heat, stand-alone devices are generally limited in terms of computing power and thus limited in the quality of content that can be rendered. The base stations of tethered systems may provide more computing power and thus higher quality rendering than stand-alone devices; however, the physical cable tethers the device to the base station and thus constrains the movements of the user.

SUMMARY

[0003] Various embodiments of methods and apparatus for providing mixed reality views to users through wireless connections are described. Embodiments of a mixed reality system are described that may include a device such as a headset, helmet, goggles, or glasses worn by the user, and a separate computing device, referred to herein as a base station. The device and base station may each include wireless communications technology that allows the device and base station to communicate and exchange data via a wireless connection. The device may include world-facing sensors that collect information about the user’s environment and user-facing sensors that collect information about the user. The information collected by the sensors may be transmitted to the base station via the wireless connection. The base station may include software and hardware configured to generate and render frames that include virtual content based at least in part on the sensor information received from the device via the wireless connection and to compress and transmit the rendered frames to the device for display via the wireless connection. The base station may provide much more computing power than can be provided by conventional stand-alone systems. In addition, the wireless connection between the device and the base station does not tether the device to the base station as in conventional tethered systems and thus allow users much more freedom of movement than do tethered systems.

[0004] Various methods and apparatus are described that may be used to maintain a target frame rate through the wireless link and to minimize latency in frame rendering, transmittal, and display.

[0005] A method that may be used in some embodiments may be referred to as warp space rendering. In the warp space rendering method, instead of performing a rectilinear projection which tends to oversample the edges of the image especially in wide FOV frames, a transform is applied that transforms the frame into a warp space. The warp space is then resampled at equal angles. The warp space rendering method resamples the frame so that rendering engine only rasterizes and renders the number of samples it actually needs no matter what direction the user is looking at. The warp space rendering method reduces the resolution of and thus the time it takes to render a frame, which reduces latency, and also reduces the number of bits that need to be transmitted over the wireless link between the device and the base station, which reduces bandwidth usage and latency.

[0006] Another method that may be used in some embodiments may be referred to as foveated rendering. In the foveated rendering method, gaze tracking information received from the device may be used to identify the direction in which the user is currently looking. A foveated region may be determined based at least in part on the determined gaze direction. Regions of the frame outside the foveated region (referred to as the peripheral region) may be converted to a lower resolution before transmission to the device, for example by applying a filter (e.g., a band pass filter) to the peripheral region. The foveated rendering method reduces the number of pixels in the rendered frame, which reduces the number of bits that need to be transmitted over the wireless link to the device, which reduces bandwidth usage and latency. In addition, in some embodiments, the peripheral region outside the foveated region of the frames may be transmitted over the wireless link at a lower frame rate than the foveated region.

[0007] Another method that may be used in some embodiments may be referred to as foveated compression. In the foveated compressing method, a foveated region and a peripheral region may be determined, either dynamically based on the gaze direction determined from gaze tracking region or statically based on a set system parameter. In some embodiments, the peripheral region may be pre-filtered to reduce information based on knowledge of the human vision system, for example by filtering high frequency information and/or increasing color compression. The amount of filtering applied to the peripheral region may increase extending towards the periphery of the image. Pre-filtering of the peripheral region may result in improved compression of the frame. Alternatively, a higher compression ratio may be used in the peripheral region than a compression ratio that is used in the foveated region.

[0008] Another method that may be used in some embodiments may be referred to as dynamic rendering. In the dynamic rendering method, to maintain a target frame rate and latency, a monitoring process on the base station monitors bandwidth on the wireless link and the rate at which the rendering application on the base station is taking to generate frames. Upon detecting that the bandwidth is below a threshold or that the frame rendering rate is below a threshold, the monitoring process may dynamically adjust one or more rendering processes on the base station to reduce the complexity of rendering a frame and thus the resolution of the rendered frames so that a target frame rate and latency to the device can be maintained. The rendering complexity may be adjusted again to increase the complexity of rendering a frame and thus increase the resolution of the frame upon detecting that the monitored metrics have reached or exceeded the threshold.

[0009] Instead of or in addition to dynamic rendering, another method that may be used in some embodiments may be referred to as dynamic compression. In the dynamic compression method, to maintain a target frame rate and latency, a monitoring process on the base station monitors bandwidth on the wireless link and the rate at which the rendering application on the base station is taking to generate frames. Upon detecting that the bandwidth is below a threshold or that the frame rendering rate is below a threshold, the monitoring process may dynamically adjust one or more compression processes on the base station to increase the compression ratio and/or increase pre-filtering of the image to reduce high frequency content so that a target frame rate and latency to the device can be maintained. The compression process(es) may be adjusted again to reduce the compression ratio and/or pre-filtering upon detecting that the monitored metrics have reached or exceeded the threshold.

[0010] Another method that may be used in some embodiments may be referred to as motion-based rendering. In this method, motion tracking information received from the device may be used to identify motion of the user’s head. If the user is not moving their head or not moving it much, frames can be rendered and sent to the device at a lower frame rate. If rapid head motion is detected, the frame rate can be increased.

[0011] Another method that may be used some embodiments may be referred to as slice-based rendering. In slice-based rendering, rather than rendering entire frames in the base station and transmitting the rendered frames to the device, the base station may render parts of frames (referred to as slices) and transmit the rendered slices to the device as they are ready. A slice may be one or more lines of a frame, or may be an N.times.M pixel section or region of a frame. Slice-based rendering reduces latency, and also reduces the amount of memory needed for buffering, which reduces the memory footprint on the chip(s) or processor(s) as well as power requirements.

[0012] In addition, methods and apparatus are described that allow the device to function as a stand-alone device as a fallback position if the wireless link with the base station is lost. In addition, methods and apparatus for processing and displaying frames received by the device from the base station via the wireless connection are described, as well as methods and apparatus for replacing incomplete or missing frames with previously received frames.

BRIEF DESCRIPTION OF THE DRAWINGS

[0013] FIG. 1 illustrates a mixed or virtual reality system, according to at least some embodiments.

[0014] FIG. 2 illustrates sensors of a device in a system as illustrated in FIG. 1, according to at least some embodiments.

[0015] FIG. 3 is a block diagram illustrating components of a mixed reality system as illustrated in FIG. 1, according to at least some embodiments.

[0016] FIG. 4 is a high-level flowchart of a method of operation for a mixed reality system as illustrated in FIGS. 1 through 3, according to at least some embodiments.

[0017] FIGS. 5A through 5D graphically illustrate warp space rendering, according to some embodiments.

[0018] FIG. 6 is a flowchart of a method for warp space rendering to reduce the resolution at which frames are rendered by the base station, according to some embodiments.

[0019] FIG. 7 graphically illustrates foveated rendering, according to some embodiments.

[0020] FIG. 8 is a flowchart of a method for foveated rendering to reduce the resolution of rendered frames before transmitting the frames over the wireless connection, according to some embodiments.

[0021] FIG. 9 is a flowchart of a method for dynamic rendering to maintain a target frame rate and latency over the wireless connection, according to some embodiments.

[0022] FIG. 10 is a flowchart of a method for motion-based rendering to maintain a target frame rate and latency over the wireless connection, according to some embodiments.

[0023] FIG. 11 is a flowchart of a method for rendering and displaying frames on the device upon detecting that the wireless connection has been lost, according to some embodiments.

[0024] FIG. 12 is a flowchart of a method for processing and displaying frames received by the device from the base station via the wireless connection, according to some embodiments.

[0025] FIG. 13 is a block diagram illustrating functional components of and processing in an example mixed reality system as illustrated in FIGS. 1 through 12, according to some embodiments.

[0026] This specification includes references to “one embodiment” or “an embodiment.” The appearances of the phrases “in one embodiment” or “in an embodiment” do not necessarily refer to the same embodiment. Particular features, structures, or characteristics may be combined in any suitable manner consistent with this disclosure.

[0027] “Comprising.” This term is open-ended. As used in the claims, this term does not foreclose additional structure or steps. Consider a claim that recites: “An apparatus comprising one or more processor units … .” Such a claim does not foreclose the apparatus from including additional components (e.g., a network interface unit, graphics circuitry, etc.).

[0028] “Configured To.” Various units, circuits, or other components may be described or claimed as “configured to” perform a task or tasks. In such contexts, “configured to” is used to connote structure by indicating that the units/circuits/components include structure (e.g., circuitry) that performs those task or tasks during operation. As such, the unit/circuit/component can be said to be configured to perform the task even when the specified unit/circuit/component is not currently operational (e.g., is not on). The units/circuits/components used with the “configured to” language include hardware–for example, circuits, memory storing program instructions executable to implement the operation, etc. Reciting that a unit/circuit/component is “configured to” perform one or more tasks is expressly intended not to invoke 35 U.S.C. .sctn. 112, paragraph (f), for that unit/circuit/component. Additionally, “configured to” can include generic structure (e.g., generic circuitry) that is manipulated by software or firmware (e.g., an FPGA or a general-purpose processor executing software) to operate in manner that is capable of performing the task(s) at issue. “Configure to” may also include adapting a manufacturing process (e.g., a semiconductor fabrication facility) to fabricate devices (e.g., integrated circuits) that are adapted to implement or perform one or more tasks.

[0029] “First,” “Second,” etc. As used herein, these terms are used as labels for nouns that they precede, and do not imply any type of ordering (e.g., spatial, temporal, logical, etc.). For example, a buffer circuit may be described herein as performing write operations for “first” and “second” values. The terms “first” and “second” do not necessarily imply that the first value must be written before the second value.

[0030] “Based On” or “Dependent On.” As used herein, these terms are used to describe one or more factors that affect a determination. These terms do not foreclose additional factors that may affect a determination. That is, a determination may be solely based on those factors or based, at least in part, on those factors. Consider the phrase “determine A based on B.” While in this case, B is a factor that affects the determination of A, such a phrase does not foreclose the determination of A from also being based on C. In other instances, A may be determined based solely on B.

[0031] “Or.” When used in the claims, the term “or” is used as an inclusive or and not as an exclusive or. For example, the phrase “at least one of x, y, or z” means any one of x, y, and z, as well as any combination thereof.

DETAILED DESCRIPTION

[0032] Various embodiments of methods and apparatus for providing mixed reality views to users through wireless connections are described. Embodiments of a mixed reality system are described that may include a device such as a headset, helmet, goggles, or glasses worn by the user, and a separate computing device, referred to herein as a base station. The device and base station may each include wireless communications technology that allows the device and base station to communicate and exchange data via a wireless connection. The device may include world-facing sensors that collect information about the user’s environment (e.g., video, depth information, lighting information, etc.), and user-facing sensors that collect information about the user (e.g., the user’s expressions, eye movement, hand gestures, etc.). The information collected by the sensors may be transmitted to the base station via the wireless connection. The base station may include software and hardware (e.g., processors (system on a chip (SOC), CPUs, image signal processors (ISPs), graphics processing units (GPUs), coder/decoders (codecs), etc.), memory, etc.) configured to generate and render frames that include virtual content based at least in part on the sensor information received from the device via the wireless connection and to compress and transmit the rendered frames to the device for display via the wireless connection.

[0033] Embodiments of the mixed reality system as described herein may collect, analyze, transfer, and store personal information, for example images of a person’s face and/or of an environment in which the person is using the system. The personal information collected by the sensors should be stored, transferred, and used only by the device and/or by the base station, and used only for the operation of the mixed reality system on the device and the base station. Embodiments will comply with well-established privacy policies and/or privacy practices. In particular, privacy policies and practices that are generally recognized as meeting or exceeding industry or governmental requirements for maintaining personal information private and secure should be implemented. For example, personal information should be collected for legitimate and reasonable uses and not shared or sold outside of those legitimate uses. Further, collection or other uses of the personal information should occur only after receiving the informed consent of the user. Additionally, any needed steps for safeguarding and securing access to such personal information and ensuring that any entity with access to the personal information adhere to the privacy policies and procedures should be taken. Further, any entity with access to the personal information can be subjected to evaluation by third parties to certify adherence to the privacy policies and practices. In addition, in some embodiments, users may selectively block the use of, or access to, their personal information. For example, hardware and/or software elements may be provided that allow a user to selectively prevent or block access to their personal information.

[0034] Conventional VR, AR, and MR systems are typically either tethered systems including a base station that performs at least some of the rendering of content for display and a device connected to the base station via a physical connection (i.e., a data communications cable), or stand-alone devices that perform rendering of content locally. Stand-alone systems allow users freedom of movement; however, because of restraints including size, weight, batteries, and heat, stand-alone devices are generally limited in terms of computing power and thus limited in the quality of content that can be rendered. The base stations of tethered systems may provide more computing power and thus higher quality rendering than stand-alone devices; however, the physical cable tethers the device to the base station and thus constrains the movements of the user.

[0035] Embodiments of the mixed reality system as described herein include a base station that provides much more computing power than can be provided by conventional stand-alone systems. In addition, the wireless connection between the device and the base station does not tether the device to the base station as in conventional tethered systems and thus allow users much more freedom of movement than do tethered systems.

[0036] In some embodiments, the mixed reality system may implement a proprietary wireless communications technology (e.g., 60 gigahertz (GHz) wireless technology) that provides a highly directional wireless link between the device and the base station. In some embodiments, the directionality and bandwidth (e.g., 60 GHz) of the wireless communication technology may support multiple devices communicating with the base station at the same time to thus enable multiple users to use the system at the same time in a co-located environment. However, other commercial (e.g., Wi-Fi, Bluetooth, etc.) or proprietary wireless communications technologies may be supported in some embodiments.

[0037] Two primary constraints to be considered on the wireless link are bandwidth and latency. A target is to provide a high resolution, wide field of view (FOV) virtual display to the user at a frame rate (e.g., 60-120 frames per second (FPS)) that provides the user with a high-quality MR view. Another target is to minimize latency between the time a video frame is captured by the device and the time a rendered MR frame based on the video frame is displayed by the device, for example to the sub-millisecond (ms) range. However, the channel capacity of the wireless link may vary with time, and the wireless link may thus support only a certain amount of information to be transmitted at any given time. Various methods and apparatus are described herein that may be used to maintain the target frame rate through the wireless link and to minimize the latency in frame rendering, transmittal, and display.

[0038] A method that may be used in some embodiments may be referred to as warp space rendering, which may be used to reduce the resolution at which frames are rendered by the base station, which reduces computation time, power usage, bandwidth usage, and latency. Ideally, there should be the same resolution on the display in any direction the user is looking. In the warp space rendering method, instead of the rendering engine of the base station performing a rectilinear projection when rendering a frame, which tends to oversample the edges of the image especially in wide FOV frames, a transform is applied that transforms the frame into a warp space. The warp space is then resampled at equal angles. The warp space rendering method resamples the frame so that rendering engine only rasterizes and renders the number of samples it actually needs no matter what direction the user is looking at. The warp space rendering method reduces the resolution of and thus the time it takes to render a frame, which reduces latency, and also reduces the number of bits that need to be transmitted over the wireless link between the device and the base station, which reduces bandwidth usage and latency.

[0039] Another method that may be used in some embodiments may be referred to as foveated rendering, which may be used to reduce the resolution of frames rendered by the base station before transmitting the frames to the device, which reduces latency and bandwidth usage. In the foveated rendering method, gaze tracking information received from the device may be used to identify the direction in which the user is currently looking. Human eyes can perceive higher resolution in the foveal region than in the peripheral region. Thus, a region of the frame that corresponds to the fovea (referred to as the foveated region) may be identified based at least in part on the determined gaze direction and transmitted to the device via the wireless connection at a higher resolution, while regions of the frame outside the foveated region (referred to as the peripheral region) may be converted to a lower resolution before transmission to the device, for example by applying a filter (e.g., a band pass filter) to the peripheral region. The foveated rendering method reduces the number of pixels in the rendered frame, which reduces the number of bits that need to be transmitted over the wireless link to the device, which reduces bandwidth usage and latency. In addition, in some embodiments, the peripheral region outside the foveated region of the frames may be transmitted over the wireless link at a lower frame rate than the foveated region.

[0040] Another method that may be used in some embodiments may be referred to as foveated compression. In the foveated compressing method, a foveated region and a peripheral region may be determined, either dynamically based on the gaze direction determined from gaze tracking region or statically based on a set system parameter. In some embodiments, the peripheral region may be pre-filtered to reduce information based on knowledge of the human vision system, for example by filtering high frequency information and/or increasing color compression. The amount of filtering applied to the peripheral region may increase extending towards the periphery of the image. Pre-filtering of the peripheral region may result in improved compression of the frame. Alternatively, a higher compression ratio may be used in the peripheral region. A tradeoff between the two methods may be either a blurrier peripheral region (through pre-filtering) or potentially higher compression artifacts (through increasing compression).

[0041] Another method that may be used in some embodiments may be referred to as dynamic rendering. In the dynamic rendering method, to maintain a target frame rate and latency, a monitoring process on the base station monitors bandwidth on the wireless link and the rate at which the rendering application on the base station is taking to generate frames. Upon detecting that the bandwidth is below a threshold or that the frame rendering rate is below a threshold, the monitoring process may dynamically adjust one or more rendering processes on the base station to reduce the complexity of rendering a frame and thus the resolution of the rendered frames so that a target frame rate and latency to the device can be maintained. The rendering complexity may be adjusted again to increase the complexity of rendering a frame and thus the resolution of the frame upon detecting that the monitored metrics have reached or exceeded the threshold.

[0042] Instead of or in addition to dynamic rendering, another method that may be used in some embodiments may be referred to as dynamic compression. In the dynamic compression method, to maintain a target frame rate and latency, a monitoring process on the base station monitors bandwidth on the wireless link and the rate at which the rendering application on the base station is taking to generate frames. Upon detecting that the bandwidth is below a threshold or that the frame rendering rate is below a threshold, the monitoring process may dynamically adjust one or more compression processes on the base station to increase the compression ratio and/or increase pre-filtering of the image to reduce high frequency content so that a target frame rate and latency to the device can be maintained. The compression process(es) may be adjusted again to reduce the compression ratio and/or pre-filtering upon detecting that the monitored metrics have reached or exceeded the threshold.

[0043] Another method that may be used in some embodiments may be referred to as motion-based rendering. In this method, motion tracking information received from the device may be used to identify motion of the user’s head. If the user is not moving their head or not moving it much, frames can be rendered and sent to the device at a lower frame rate. There may be little or no perceived difference to the user at the lower frame rate because the user’s head is not in rapid motion. If rapid head motion is detected, the frame rate can be increased.

[0044] Another method that may be used some embodiments may be referred to as slice-based rendering. Rendering and transmitting entire frames may have a latency and memory impact as each frame needs to be completed, stored, and then transmitted to the next stage of the mixed reality system. In slice-based rendering, rather than rendering entire frames in the base station and transmitting the rendered frames to the device, the base station may render parts of frames (referred to as slices) and transmit the rendered slices to the device as they are ready. A slice may be one or more lines of a frame, or may be an N.times.M pixel section or region of a frame. Slice-based rendering reduces latency, and also reduces the amount of memory needed for buffering, which reduces the memory footprint on the chip(s) or processor(s) as well as power requirements. Note that the term “frame portion” may be used herein to refer to an entire frame or to a slice of a frame as described above.

[0045] In addition, methods and apparatus are described that allow the device to function as a stand-alone device as a fallback position if the wireless link with the base station is lost, for example if the base station goes down or an object comes between the device and the base station, blocking the wireless link.

[0046] FIG. 1 illustrates a mixed or virtual reality system 10, according to at least some embodiments. In some embodiments, a system 10 may include a device 100, and a base station 160 configured to render mixed reality frames including virtual content 110 for display by the device 100. Device 100 may, for example be a head-mounted device (HMD) such as a headset, helmet, goggles, or glasses that may be worn by a user 190. The mixed reality frames may include computer generated information (referred to as virtual content) composited with real world images or a real world view to augment, or add content to, a user’s view of the world, or alternatively may include representations of real world objects composited with views of a computer generated three-dimensional (3D) virtual world. The device 100 and base station 160 may each include wireless communications technology that allows the device 100 and base station 160 to communicate and exchange data via a wireless connection 180.

[0047] The device 100 may include sensors 140 and 150 that collect information about the user 190’s environment (video, depth information, lighting information, etc.), and information about the user 190 (e.g., the user’s expressions, eye movement, gaze direction, hand gestures, etc.). Example sensors 140 and 150 are shown in FIG. 2. The device 100 may transmit at least some of the information collected by sensors 140 and 150 to a base station 160 of the system 10 via a wireless connection 180. The base station 160 may render frames for display by the device 100 that include virtual content 110 based at least in part on the various information obtained from the sensors 140 and 150, compress the frames, and transmit the frames to the device 100 for display to the user 190 via the wireless connection 180. The information collected by the sensors 140 and 150 should be stored, transferred, and used only by the device 100 and/or by the base station 160, and used only for the operation of the mixed reality system on the device 100 and the base station 160.

[0048] A 3D virtual view 102 may be a three-dimensional (3D) space including virtual content 110 at different depths that a user 190 sees when using a mixed or virtual reality system 10. In some embodiments, virtual content 110 may be displayed to the user 190 in the 3D virtual view 102 by the device 100; in the 3D virtual view 102, different virtual objects may be displayed at different depths in a 3D virtual space. In some embodiments, in the 3D virtual view 102, the virtual content 110 may be overlaid on or composited in a view of the user 190’s environment with respect to the user’s current line of sight that is provided by the device 100. Device 100 may implement any of various types of virtual reality projection technologies. For example, device 100 may be a near-eye VR system that displays left and right images on screens in front of the user 190’s eyes that are viewed by a subject, such as DLP (digital light processing), LCD (liquid crystal display) and LCoS (liquid crystal on silicon) technology VR systems. In some embodiments, the screens may be see-through displays. As another example, device 100 may be a direct retinal projector system that scans left and right images, pixel by pixel, to the subject’s eyes. To scan the images, left and right projectors generate beams that are directed to left and right reflective components (e.g., ellipsoid mirrors) located in front of the user 190’s eyes; the reflective components reflect the beams to the user’s eyes. To create a three-dimensional (3D) effect, virtual content 110 at different depths or distances in the 3D virtual view 102 are shifted left or right in the two images as a function of the triangulation of distance, with nearer objects shifted more than more distant objects.

[0049] While not shown in FIG. 1, in some embodiments the mixed reality system 10 may include one or more other components. For example, the system may include a cursor control device (e.g., mouse) for moving a virtual cursor in the 3D virtual view 102 to interact with virtual content 110.

[0050] While FIG. 1 shows a single user 190 and device 100, in some embodiments the mixed reality system 10 may support multiple devices 100 communicating with the base station 160 at the same time to thus enable multiple users 190 to use the system at the same time in a co-located environment.

[0051] FIG. 2 illustrates sensors of an example device 200, according to at least some embodiments. FIG. 2 shows a side view of an example device 200 with sensors, according to some embodiments. Note that device 200 as illustrated in FIG. 2 is given by way of example, and is not intended to be limiting. In various embodiments, the shape, size, and other features of the device may differ, and the locations, numbers, types, and other features of the world and user sensors may vary. The device 200 may, for example, be a head-mounted device (HMD) such as a headset, helmet, goggles, or glasses worn by the user.

[0052] The device 200 may include sensors that collect information about the user 290’s environment (video, depth information, lighting information, etc.) and information about the user 290 (e.g., the user’s expressions, eye movement, hand gestures, etc.). In some embodiments, the device 200 may be worn by a user 290’s so that the projection system displays 202 (e.g. screens and optics of a near-eye VR system, or reflective components (e.g., ellipsoid mirrors) of a direct retinal projector system) are disposed in front of the user 290’s eyes 292.

[0053] The device 200 may include one or more of various types of processors 204 (system on a chip (SOC), CPUs, image signal processors (ISPs), graphics processing units (GPUs), coder/decoders (codecs), etc.) that may, for example perform initial processing (e.g., compression) of the information collected by the sensors and transmit the information to a base station 260 of the mixed reality system via a wireless connection 280, and that may also perform processing (e.g., decoding/decompression) of compressed frames received from the base station 260 and provide the processed frames to the display subsystem for display. In some embodiments, virtual content may be displayed to the user 290 in a 3D virtual view by the device 200; in the 3D virtual view, different virtual objects may be displayed at different depths in a 3D virtual space. In some embodiments, in the 3D virtual view, the virtual content may be overlaid on or composited in a view of the user 290’s environment with respect to the user’s current line of sight that is provided by the device 200.

[0054] In some embodiments, the wireless connection 280 may be implemented according to a proprietary wireless communications technology (e.g., 60 gigahertz (GHz) wireless technology) that provides a highly directional wireless link between the device 200 and the base station 260. However, other commercial (e.g., Wi-Fi, Bluetooth, etc.) or proprietary wireless communications technologies may be used in some embodiments.

[0055] The base station 260 may be an external device (e.g., a computing system, game console, etc.) that is communicatively coupled to device 200 via a wireless interface. The base station 260 may include one or more of various types of processors 262 (e.g., SOCs, CPUs, ISPs, GPUs, codecs, and/or other components for processing and rendering video and/or images). The base station 260 may render frames (each frame including a left and right image) that include virtual content based at least in part on the various inputs obtained from the sensors via the wireless connection 280, compress the rendered frames, and transmit the compressed frames to the device 200 for display to the left and right displays 202. FIGS. 3 and 12 further illustrate components and operations of a device 200 and base station 260 of a mixed reality system, according to some embodiments.

[0056] Device sensors may, for example, be located on external and internal surfaces of a device 200, and may collect various information about the user 290 and about the user’s environment. In some embodiments, the information collected by the sensors may be used to provide the user with a virtual view of their real environment. The information collected by the sensors should be stored, transferred, and used only by the device 200 and/or by the base station 260, and used only for the operation of the mixed reality system on the device 200 and the base station 260. In some embodiments, the sensors may be used to provide depth information for objects in the real environment. In some embodiments, the sensors may be used to provide orientation and motion information for the user in the real environment. In some embodiments, the sensors may be used to collect color and lighting information in the real environment. In some embodiments, the information collected by the sensors may be used to adjust the rendering of images to be projected, and/or to adjust the projection of the images by the projection system of the device 200. In some embodiments, the information collected by the sensors may be used in generating an avatar of the user 290 in the 3D virtual view projected to the user by the device 200. In some embodiments, the information collected by the sensors may be used in interacting with or manipulating virtual content in the 3D virtual view projected by the device 200. In some embodiments, the user information collected by one or more user-facing sensors may be used to adjust the collection of, and/or processing of information collected by one or more world-facing sensors.

[0057] In some embodiments, the sensors may include one or more scene cameras 220 (e.g., RGB (visible light) video cameras) that capture high-quality video of the user’s environment that may be used to provide the user 290 with a virtual view of their real environment. In some embodiments, video streams captured by cameras 220 may be compressed by the device 200 and transmitted to the base station 260 via wireless connection 280. The frames may be decompressed and processed by the base station 260 at least in part according to other sensor information received from the device 200 via the wireless connection 280 to render frames including virtual content; the rendered frames may then be compressed and transmitted to the device 200 via the wireless connection 280 for display to the user 290.

[0058] In some embodiments, if the wireless connection 280 to the base station 200 is lost for some reason, at least some video frames captured by cameras 200 may be processed by processors 204 of device 200 to provide a virtual view of the real environment to the user 290 via display 202. This may, for example, be done for safety reasons so that the user 290 can still view the real environment that they are in even if the base station 260 is unavailable. In some embodiments, the processors 204 may render virtual content to be displayed in the virtual view, for example a message informing the user 290 that the wireless connection 280 has been lost.

[0059] In an example non-limiting embodiment, scene cameras 220 may include high quality, high resolution RGB video cameras, for example 10 megapixel (e.g., 3072.times.3072 pixel count) cameras with a frame rate of 60 frames per second (FPS) or greater, horizontal field of view (HFOV) of greater than 90 degrees, and with a working distance of 0.1 meters (m) to infinity. In some embodiments there may be two scene cameras 220 (e.g., a left and a right camera 220) located on a front surface of the device 200 at positions that are substantially in front of each of the user 290’s eyes 292. However, more or fewer scene cameras 220 may be used in a device 200 to capture video of the user 290’s environment, and scene cameras 220 may be positioned at other locations.

[0060] In some embodiments, the sensors may include one or more world mapping sensors (e.g., infrared (IR) cameras with an IR illumination source, or Light Detection and Ranging (LIDAR) emitters and receivers/detectors) that, for example, capture depth or range information for objects and surfaces in the user’s environment. The range information may, for example, be used in positioning virtual content composited with images of the real environment at correct depths. In some embodiments, the range information may be used in adjusting the depth of real objects in the environment when displayed; for example, nearby objects may be re-rendered to be smaller in the display to help the user in avoiding the objects when moving about in the environment. In some embodiments there may be one world mapping sensor located on a front surface of the device 200. However, in various embodiments, more than one world mapping sensor may be used, and world mapping sensor(s) may be positioned at other locations. In an example non-limiting embodiment, a world mapping sensor may include an IR light source and IR camera, for example a 1 megapixel (e.g., 1000.times.1000 pixel count) camera with a frame rate of 60 frames per second (FPS) or greater, HFOV of 90 degrees or greater, and with a working distance of 0.1 m to 1.5 m.

[0061] In some embodiments, the sensors may include one or more head pose sensors (e.g., IR or RGB cameras) that may capture information about the position, orientation, and/or motion of the user and/or the user’s head in the environment. The information collected by head pose sensors may, for example, be used to augment information collected by an inertial-measurement unit (IMU) 206 of the device 200. The augmented position, orientation, and/or motion information may be used in determining how to render and display virtual views of the user’s environment and virtual content within the views. For example, different views of the environment may be rendered based at least in part on the position or orientation of the user’s head, whether the user is currently walking through the environment, and so on. As another example, the augmented position, orientation, and/or motion information may be used to composite virtual content with the scene in a fixed position relative to the background view of the user’s environment. In some embodiments there may be two head pose sensors located on a front or top surface of the device 200. However, in various embodiments, more or fewer head pose sensors may be used, and the sensors may be positioned at other locations. In an example non-limiting embodiment, head pose sensors may include RGB or IR cameras, for example 400.times.400 pixel count cameras, with a frame rate of 120 frames per second (FPS) or greater, wide field of view (FOV), and with a working distance of 1 m to infinity. The head pose sensors may include wide FOV lenses, and may look in different directions. The head pose sensors may provide low latency monochrome imaging for tracking head position and motion, and may be integrated with an IMU of the device 200 to augment head position and movement information captured by the IMU.

[0062] In some embodiments, the sensors may include one or more light sensors (e.g., RGB cameras) that capture lighting information (e.g., direction, color, and intensity) in the user’s environment that may, for example, be used in rendering virtual content in the virtual view of the user’s environment, for example in determining coloring, lighting, shadow effects, etc. for virtual objects in the virtual view. For example, if a red light source is detected, virtual content rendered into the scene may be illuminated with red light, and more generally virtual objects may be rendered with light of a correct color and intensity from a correct direction and angle. In some embodiments there may be one light sensor located on a front or top surface of the device 200. However, in various embodiments, more than one light sensor may be used, and light sensor(s) may be positioned at other locations. In an example non-limiting embodiment, a light sensor may include an RGB high dynamic range (HDR) video camera, for example a 500.times.500 pixel count camera, with a frame rate of 30 FPS, HFOV of 180 degrees or greater, and with a working distance of 1 m to infinity.

[0063] In some embodiments, the sensors may include one or more gaze tracking sensors 224 (e.g., IR cameras with an IR illumination source) that may be used to track position and movement of the user’s eyes. In some embodiments, gaze tracking sensors 224 may also be used to track dilation of the user’s pupils. In some embodiments, there may be two gaze tracking sensors 224, with each gaze tracking sensor tracking a respective eye 292. In some embodiments, the information collected by the gaze tracking sensors 224 may be used to adjust the rendering of images to be projected, and/or to adjust the projection of the images by the projection system of the device 200, based on the direction and angle at which the user’s eyes are looking. For example, in some embodiments, content of the images in a region around the location at which the user’s eyes are currently looking may be rendered with more detail and at a higher resolution than content in regions at which the user is not looking, which allows available processing time for image data to be spent on content viewed by the foveal regions of the eyes rather than on content viewed by the peripheral regions of the eyes. Similarly, content of images in regions at which the user is not looking may be compressed more than content of the region around the point at which the user is currently looking. In some embodiments, the information collected by the gaze tracking sensors 224 may be used to match direction of the eyes of an avatar of the user 290 to the direction of the user’s eyes. In some embodiments, brightness of the projected images may be modulated based on the user’s pupil dilation as determined by the gaze tracking sensors 224. In some embodiments there may be two gaze tracking sensors 224 located on an inner surface of the device 200 at positions such that the sensors 224 have views of respective ones of the user 290’s eyes 292. However, in various embodiments, more or fewer gaze tracking sensors 224 may be used in a device 200, and sensors 224 may be positioned at other locations. In an example non-limiting embodiment, each gaze tracking sensor 224 may include an IR light source and IR camera, for example a 400.times.400 pixel count camera with a frame rate of 120 FPS or greater, HFOV of 70 degrees, and with a working distance of 10 millimeters (mm) to 80 mm.

[0064] In some embodiments, the sensors may include one or more sensors (e.g., IR cameras with IR illumination) that track expressions of the user’s forehead area and/or of the user’s mouth/jaw area. In some embodiments, expressions of the brow, mouth, jaw, and eyes captured by the user-facing sensors may be used to simulate expressions on an avatar in the virtual space, and/or to selectively render and composite virtual content based at least in part on the user’s reactions to projected content. In some embodiments there may be two sensors located on an inner surface of the device 200 at positions such that the sensors have views of the user 290’s forehead, and two sensors located on an inner surface of the device 200 at positions such that the sensors have views of the user 290’s lower jaw and mouth. However, in various embodiments, more or fewer sensors may be used in a device 200, and the sensors may be positioned at other locations than those shown. In an example non-limiting embodiment, each sensor may include an IR light source and IR camera. In some embodiments, images from two or more of the sensors may be combined to form a stereo view of a portion of the user’s faces.

[0065] In some embodiments, the sensors may include one or more sensors (e.g., IR cameras with IR illumination) that track position, movement, and gestures of the user’s hands, fingers, and/or arms. As an example, the user’s detected hand and finger gestures may be used to determine interactions of the user with virtual content in the virtual space, including but not limited to gestures that manipulate virtual objects, gestures that interact with virtual user interface elements displayed in the virtual space, etc. In some embodiments there may be one sensor located on a bottom surface of the device 200. However, in various embodiments, more than one sensor may be used, and sensors may be positioned at other locations. In an example non-limiting embodiment, a sensor may include an IR light source and IR camera.

[0066] FIG. 3 is a block diagram illustrating components of an example mixed reality system, according to at least some embodiments. In some embodiments, a mixed reality system may include a device 300 and a base station 360 (e.g., a computing system, game console, etc.). The device 300 may, for example, be a head-mounted device (HMD) such as a headset, helmet, goggles, or glasses worn by the user.

[0067] Device 300 may include a display 302 component or subsystem that may implement any of various types of virtual reality projector technologies. For example, the device 300 may include a near-eye VR projector that displays frames including left and right images on screens that are viewed by a user, such as DLP (digital light processing), LCD (liquid crystal display) and LCoS (liquid crystal on silicon) technology projectors. In some embodiments, the screens may be see-through displays. As another example, the device 300 may include a direct retinal projector that scans frames including left and right images, pixel by pixel, directly to the user’s eyes via a reflective surface (e.g., reflective eyeglass lenses). To create a three-dimensional (3D) effect in 3D virtual view 310, objects at different depths or distances in the two images are shifted left or right as a function of the triangulation of distance, with nearer objects shifted more than more distant objects.

[0068] The device 300 may also include a controller 304 configured to implement device-side functionality of the mixed reality system as described herein. In some embodiments, device 300 may also include a memory 330 configured to store software (code 332) of the device component of the mixed reality system that is executable by the controller 304, as well as data 334 that may be used by the code 332 when executing on the controller 304.

[0069] In various embodiments, controller 304 may be a uniprocessor system including one processor, or a multiprocessor system including several processors (e.g., two, four, eight, or another suitable number). Controller 304 may include central processing units (CPUs) configured to implement any suitable instruction set architecture, and may be configured to execute instructions defined in that instruction set architecture. For example, in various embodiments controller 304 may include general-purpose or embedded processors implementing any of a variety of instruction set architectures (ISAs), such as the x86, PowerPC, SPARC, RISC, or MIPS ISAs, or any other suitable ISA. In multiprocessor systems, each of the processors may commonly, but not necessarily, implement the same ISA. Controller 304 may employ any microarchitecture, including scalar, superscalar, pipelined, superpipelined, out of order, in order, speculative, non-speculative, etc., or combinations thereof. Controller 304 may include circuitry to implement microcoding techniques. Controller 304 may include one or more processing cores each configured to execute instructions. Controller 304 may include one or more levels of caches, which may employ any size and any configuration (set associative, direct mapped, etc.).

[0070] In some embodiments, controller 304 may include at least one graphics processing unit (GPU), which may include any suitable graphics processing circuitry. Generally, a GPU may be configured to render objects to be displayed into a frame buffer (e.g., one that includes pixel data for an entire frame). A GPU may include one or more graphics processors that may execute graphics software to perform a part or all of the graphics operation, or hardware acceleration of certain graphics operations. In some embodiments, controller 304 may include one or more other components for processing and rendering video and/or images, for example image signal processors (ISPs), coder/decoders (codecs), etc. In some embodiments, controller 304 may include at least one system on a chip (SOC).

[0071] Memory 330 may include any type of memory, such as dynamic random access memory (DRAM), synchronous DRAM (SDRAM), double data rate (DDR, DDR2, DDR3, etc.) SDRAM (including mobile versions of the SDRAMs such as mDDR3,etc., or low power versions of the SDRAMs such as LPDDR2, etc.), RAMBUS DRAM (RDRAM), static RAM (SRAM), etc., or memory modules such as single inline memory modules (SIMMs), dual inline memory modules (DIMMs), etc. In some embodiments, memory devices may be mounted with an integrated circuit implementing system in a chip-on-chip configuration, a package-on-package configuration, or a multi-chip module configuration.

[0072] In some embodiments, the device 300 may include sensors of various types. In some embodiments, the device 300 may include at least one inertial-measurement unit (IMU) 306 configured to detect position, orientation, and/or motion of the device 300, and to provide the detected position, orientation, and/or motion data to the controller 304 of the device 300. In some embodiments, the device 300 may include sensors 320 and 322 that collect information about the user’s environment (video, depth information, lighting information, etc.) and about the user (e.g., the user’s expressions, eye movement, hand gestures, etc.). The sensors 320 and 322 may provide the collected information to the controller 304 of the device 300. Sensors 320 and 322 may include, but are not limited to, visible light cameras (e.g., video cameras), infrared (IR) cameras, IR cameras with an IR illumination source, Light Detection and Ranging (LIDAR) emitters and receivers/detectors, and laser-based sensors with laser emitters and receivers/detectors. Sensors of an example device are shown in FIG. 2.

[0073] The device 300 may also include one or more wireless technology interfaces 308 configured to communicate with an external base station 360 via a wireless connection 380 to send sensor inputs to the base station 360 and receive compressed rendered frames or slices from the base station 360. In some embodiments, a wireless technology interface 308 may implement a proprietary wireless communications technology (e.g., 60 gigahertz (GHz) wireless technology) that provides a highly directional wireless link between the device 300 and the base station 360. However, other commercial (e.g., Wi-Fi, Bluetooth, etc.) or proprietary wireless communications technologies may be used in some embodiments.

[0074] Base station 360 may be or may include any type of computing system or computing device, such as a desktop computer, notebook or laptop computer, pad or tablet device, smartphone, hand-held computing device, game controller, game system, and so on. Base station 360 may include a controller 362 comprising one or more processors configured to implement base-side functionality of the mixed reality system as described herein. Base station 360 may also include a memory 364 configured to store software (code 366) of the base station component of the mixed reality system that is executable by the controller 362, as well as data 368 that may be used by the code 366 when executing on the controller 362.

[0075] In various embodiments, controller 362 may be a uniprocessor system including one processor, or a multiprocessor system including several processors (e.g., two, four, eight, or another suitable number). Controller 362 may include central processing units (CPUs) configured to implement any suitable instruction set architecture, and may be configured to execute instructions defined in that instruction set architecture. For example, in various embodiments controller 362 may include general-purpose or embedded processors implementing any of a variety of instruction set architectures (ISAs), such as the x86, PowerPC, SPARC, RISC, or MIPS ISAs, or any other suitable ISA. In multiprocessor systems, each of the processors may commonly, but not necessarily, implement the same ISA. Controller 362 may employ any microarchitecture, including scalar, superscalar, pipelined, superpipelined, out of order, in order, speculative, non-speculative, etc., or combinations thereof. Controller 362 may include circuitry to implement microcoding techniques. Controller 362 may include one or more processing cores each configured to execute instructions. Controller 362 may include one or more levels of caches, which may employ any size and any configuration (set associative, direct mapped, etc.).

[0076] In some embodiments, controller 362 may include at least one graphics processing unit (GPU), which may include any suitable graphics processing circuitry. Generally, a GPU may be configured to render objects to be displayed into a frame buffer (e.g., one that includes pixel data for an entire frame). A GPU may include one or more graphics processors that may execute graphics software to perform a part or all of the graphics operation, or hardware acceleration of certain graphics operations. In some embodiments, controller 362 may include one or more other components for processing and rendering video and/or images, for example image signal processors (ISPs), coder/decoders (codecs), etc. In some embodiments, controller 362 may include at least one system on a chip (SOC).

[0077] Memory 364 may include any type of memory, such as dynamic random access memory (DRAM), synchronous DRAM (SDRAM), double data rate (DDR, DDR2, DDR3, etc.) SDRAM (including mobile versions of the SDRAMs such as mDDR3, etc., or low power versions of the SDRAMs such as LPDDR2, etc.), RAMBUS DRAM (RDRAM), static RAM (SRAM), etc., or memory modules such as single inline memory modules (SIMMs), dual inline memory modules (DIMMs), etc. In some embodiments, memory devices may be mounted with an integrated circuit implementing system in a chip-on-chip configuration, a package-on-package configuration, or a multi-chip module configuration.

[0078] Base station 360 may also include one or more wireless technology interfaces 370 configured to communicate with device 300 via a wireless connection 380 to receive sensor inputs from the device 300 and send compressed rendered frames or slices from the base station 360 to the device 300. In some embodiments, a wireless technology interface 370 may implement a proprietary wireless communications technology (e.g., 60 gigahertz (GHz) wireless technology) that provides a highly directional wireless link between the device 300 and the base station 360. In some embodiments, the directionality and band width (e.g., 60 GHz) of the wireless communication technology may support multiple devices 300 communicating with the base station 360 at the same time to thus enable multiple users to use the system at the same time in a co-located environment. However, other commercial (e.g., Wi-Fi, Bluetooth, etc.) or proprietary wireless communications technologies may be used in some embodiments.

[0079] The base station 360 may be configured to render and transmit frames to the device 300 to provide a 3D virtual view 310 for the user based at least in part on world sensor 320 and user sensor 322 inputs received from the device 300. The virtual view 310 may include renderings of the user’s environment, including renderings of real objects 312 in the user’s environment, based on video captured by one or more scene cameras (e.g., RGB (visible light) video cameras) that capture high-quality, high-resolution video of the user’s environment in real time for display. The virtual view 310 may also include virtual content (e.g., virtual objects, 314, virtual tags 315 for real objects 312, avatars of the user, etc.) rendered and composited with the projected 3D view of the user’s real environment by the base station 360. FIG. 4 describes an example method for collecting and processing sensor inputs to generate content in a 3D virtual view 310 that may be used in a mixed reality system as illustrated in FIG. 3, according to some embodiments.

[0080] FIG. 4 is a high-level flowchart of a method of operation for a mixed reality system as illustrated in FIGS. 1 through 3, according to at least some embodiments. The mixed reality system may include a device such as a headset, helmet, goggles, or glasses that includes a display component for displaying frames including left and right images to a user’s eyes to thus provide 3D virtual views to the user. The 3D virtual views may include views of the user’s environment augmented with virtual content (e.g., virtual objects, virtual tags, etc.). The mixed reality system may also include a base station configured to receive sensor inputs, including frames captured by cameras on the device as well as eye and motion tracking inputs, from the device via a wireless interface, render mixed reality frames at least in part according to the sensor inputs, compress the mixed reality frames, and transmit the compressed frames to the device via the wireless interface for display.

[0081] As indicated at 400, one or more world sensors on the device may capture information about the user’s environment (e.g., video, depth information, lighting information, etc.), and provide the information as inputs to a controller of the device. As indicated at 410, one or more user sensors on the device may capture information about the user (e.g., the user’s expressions, eye movement, hand gestures, etc.), and provide the information as inputs to the controller of the device. Elements 410 and 420 may be performed in parallel, and may be performed continuously to provide sensor inputs as the user uses the mixed reality system. As indicated at 420, the device sends at least some of the sensor data to the base station over the wireless connection. In some embodiments, the controller of the device may perform some processing of the sensor data, for example compression, before transmitting the sensor data to the base station. As indicated at 430, the controller of the base station may render frame portions (a frame portion may include an entire frame or a slice of a frame) including virtual content based at least in part on the inputs from the world and user sensors received from the device via the wireless connection. As indicated at 440, the base station compresses the rendered frames or slices and sends the compressed frames or slices to the device over the wireless connection. As indicated at 450, the device decompresses the frames or slices received from the base station and displays the frames or slices to provide a 3D virtual view including the virtual content and a view of the user’s environment for viewing by the user. As indicated by the arrow returning from element 460 to element 400, the base station may continue to receive and process inputs from the sensors to render frames or slices for display by the device as long as the user is using the mixed reality system.

[0082] Rendering and transmitting entire frames may have a latency and memory impact as each frame needs to be completed, stored, and then transmitted to the next stage of the mixed reality system. In some embodiments, rather than rendering entire frames in the base station and transmitting the rendered frames to the device, the base station may render parts of frames (referred to as slices) and transmit the rendered slices to the device as they are ready. A slice may be one or more lines of a frame, or may be an N.times.M pixel section or region of a frame. Note that the term “frame portion” may be used herein to refer to an entire frame or to a slice of a frame as described above.

Bandwidth and Latency Constraints on the Wireless Connection

[0083] Two primary constraints to be considered on the wireless link between the device and the base station in a mixed reality system as illustrated in FIGS. 1 through 4 are bandwidth and latency. A target is to provide a high resolution, wide field of view (FOV) virtual display to the user at a frame rate (e.g., 60-120 frames per second (FPS)) that provides the user with a high-quality MR view. Another target is to minimize latency between the time a video frame is captured by the device and the time a rendered MR frame based on the video frame is displayed by the device, for example to the sub-millisecond (ms) range. Various methods and apparatus may be used in embodiments to maintain the target frame rate through the wireless link and to minimize the latency in frame rendering, transmittal, and display.

Warp Space Rendering

[0084] Some embodiments may employ warp space rendering to reduce the resolution of frames captured by the scene cameras, which reduces computation time, power usage, bandwidth usage, and latency. Ideally, there should be the same resolution on the display in any direction the user is looking. In the warp space rendering method, instead of performing a rectilinear projection when rendering a frame, which tends to oversample the edges of the image especially in wide FOV frames, a transform is applied that transforms the frame into a warp space. The warp space is then resampled at equal angles. The resampling at equal angles results in a warp space frame that has lower resolution towards the edges, and the rendering process when applied to the warp space frame results in a rendered frame that provides the same or similar resolution on the display in any direction the user is looking.

[0085] FIGS. 5A through 5D graphically illustrate warp space rendering, according to some embodiments. FIGS. 5A and 5B illustrate conventional rectilinear rendering. FIG. 5A illustrates firing rays from a view point to sample a frame 500 using a conventional rectilinear projection method. In the rectilinear projection method, rays are fired from a view point into a 3D virtual space at an equal distance d to resample a frame 500. The resampled frame is then rendered by the rendering application to generate an image for display.

[0086] As shown in FIG. 5B, the rectilinear projection method generates an image with the same resolution 504 across the display 502. Distance d may be selected to provide good detail when the user is looking at the center of the display 502. However, the human eye 592 can only resolve detail at a certain angular resolution 594. As can be seen in FIG. 5B, when the user is looking towards the edges of the display 502, the image resolution is higher than the eye’s angular resolution 594. Thus, the rectilinear projection method tends to oversample towards the edges of the image. This is especially true for wide field of view displays.

[0087] FIGS. 5C and 5D illustrate warp space rendering. FIG. 5C illustrates firing rays from a view point to sample a frame 500 using a warp space projection method. In the warp space projection method, the frame 500 is transformed into a warp space 510, and rays are fired from a view point into a 3D virtual space at an equal angle A to resample a frame 500. The resampled frame is then rendered by the rendering application to generate an image for display.

发表评论

电子邮件地址不会被公开。 必填项已用*标注